Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
30,030
11,790,540,241
IssuesEvent
2020-03-17 19:10:17
kaum-software/kaum-forum
https://api.github.com/repos/kaum-software/kaum-forum
opened
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js
security vulnerability
## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/kaum-forum/node_modules/sockjs/examples/hapi/html/index.html</p> <p>Path to vulnerable library: /kaum-forum/node_modules/sockjs/examples/hapi/html/index.html,/kaum-forum/node_modules/sockjs/examples/express/index.html,/kaum-forum/node_modules/sockjs/examples/echo/index.html,/kaum-forum/node_modules/sockjs/examples/multiplex/index.html,/kaum-forum/node_modules/sockjs/examples/express-3.x/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kaum-software/kaum-forum/commit/46f95cabd497e8f42a4afe4e5afd5c9438b9bb9a">46f95cabd497e8f42a4afe4e5afd5c9438b9bb9a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/kaum-forum/node_modules/sockjs/examples/hapi/html/index.html</p> <p>Path to vulnerable library: /kaum-forum/node_modules/sockjs/examples/hapi/html/index.html,/kaum-forum/node_modules/sockjs/examples/express/index.html,/kaum-forum/node_modules/sockjs/examples/echo/index.html,/kaum-forum/node_modules/sockjs/examples/multiplex/index.html,/kaum-forum/node_modules/sockjs/examples/express-3.x/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kaum-software/kaum-forum/commit/46f95cabd497e8f42a4afe4e5afd5c9438b9bb9a">46f95cabd497e8f42a4afe4e5afd5c9438b9bb9a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm kaum forum node modules sockjs examples hapi html index html path to vulnerable library kaum forum node modules sockjs examples hapi html index html kaum forum node modules sockjs examples express index html kaum forum node modules sockjs examples echo index html kaum forum node modules sockjs examples multiplex index html kaum forum node modules sockjs examples express x index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
15,842
20,028,188,823
IssuesEvent
2022-02-02 00:26:53
googleapis/java-translate
https://api.github.com/repos/googleapis/java-translate
closed
com.example.translate.BatchTranslateTextWithGlossaryAndModelTests: testBatchTranslateTextWithGlossaryAndModel failed
type: process api: translate priority: p3 flakybot: issue flakybot: flaky
Note: #500 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b2842d1ea50d485512b991adf4c1f6f20f3bed89 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/795135e9-d1f0-4b7d-8464-7ef3c1cd7b31), [Sponge](http://sponge2/795135e9-d1f0-4b7d-8464-7ef3c1cd7b31) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_WITH_MODEL_OUTPUT/937869c4-f826-4498-a64a-77f12d70354f/ at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:566) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:445) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:93) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:68) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.translate.BatchTranslateTextWithGlossaryAndModel.batchTranslateTextWithGlossaryAndModel(BatchTranslateTextWithGlossaryAndModel.java:117) at com.example.translate.BatchTranslateTextWithGlossaryAndModelTests.testBatchTranslateTextWithGlossaryAndModel(BatchTranslateTextWithGlossaryAndModelTests.java:120) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at com.google.cloud.testing.junit4.MultipleAttemptsRule$1.evaluate(MultipleAttemptsRule.java:94) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_WITH_MODEL_OUTPUT/937869c4-f826-4498-a64a-77f12d70354f/ at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:49) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1041) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1215) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:553) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:68) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:739) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:718) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_WITH_MODEL_OUTPUT/937869c4-f826-4498-a64a-77f12d70354f/ at io.grpc.Status.asRuntimeException(Status.java:535) ... 17 more </pre></details>
1.0
com.example.translate.BatchTranslateTextWithGlossaryAndModelTests: testBatchTranslateTextWithGlossaryAndModel failed - Note: #500 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b2842d1ea50d485512b991adf4c1f6f20f3bed89 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/795135e9-d1f0-4b7d-8464-7ef3c1cd7b31), [Sponge](http://sponge2/795135e9-d1f0-4b7d-8464-7ef3c1cd7b31) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_WITH_MODEL_OUTPUT/937869c4-f826-4498-a64a-77f12d70354f/ at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:566) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:445) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:93) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:68) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.translate.BatchTranslateTextWithGlossaryAndModel.batchTranslateTextWithGlossaryAndModel(BatchTranslateTextWithGlossaryAndModel.java:117) at com.example.translate.BatchTranslateTextWithGlossaryAndModelTests.testBatchTranslateTextWithGlossaryAndModel(BatchTranslateTextWithGlossaryAndModelTests.java:120) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at com.google.cloud.testing.junit4.MultipleAttemptsRule$1.evaluate(MultipleAttemptsRule.java:94) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_WITH_MODEL_OUTPUT/937869c4-f826-4498-a64a-77f12d70354f/ at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:49) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1041) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1215) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:553) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:68) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:739) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:718) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_WITH_MODEL_OUTPUT/937869c4-f826-4498-a64a-77f12d70354f/ at io.grpc.Status.asRuntimeException(Status.java:535) ... 17 more </pre></details>
process
com example translate batchtranslatetextwithglossaryandmodeltests testbatchtranslatetextwithglossaryandmodel failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output java util concurrent executionexception com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation with model output at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com example translate batchtranslatetextwithglossaryandmodel batchtranslatetextwithglossaryandmodel batchtranslatetextwithglossaryandmodel java at com example translate batchtranslatetextwithglossaryandmodeltests testbatchtranslatetextwithglossaryandmodel batchtranslatetextwithglossaryandmodeltests java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at com google cloud testing multipleattemptsrule evaluate multipleattemptsrule java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation with model output at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation with model output at io grpc status asruntimeexception status java more
1
16,284
20,907,402,079
IssuesEvent
2022-03-24 04:49:12
javaer996/javaer996-comments
https://api.github.com/repos/javaer996/javaer996-comments
opened
Spring系列-BeanPostProcessor详解 - TENG JIANG BLOG
Gitalk /Spring%E7%B3%BB%E5%88%97-BeanPostProcessor.html
https://www.tengjiang.site/Spring%E7%B3%BB%E5%88%97-BeanPostProcessor.html BeanFactoryPostProcessor MergedBeanDefinitionPostProcessor InstantiationAwareBeanPostProcessor SmartInstantiationAwareBeanPostProcessor DestructionAwareBeanPostProcessor
1.0
Spring系列-BeanPostProcessor详解 - TENG JIANG BLOG - https://www.tengjiang.site/Spring%E7%B3%BB%E5%88%97-BeanPostProcessor.html BeanFactoryPostProcessor MergedBeanDefinitionPostProcessor InstantiationAwareBeanPostProcessor SmartInstantiationAwareBeanPostProcessor DestructionAwareBeanPostProcessor
process
spring系列 beanpostprocessor详解 teng jiang blog beanfactorypostprocessor mergedbeandefinitionpostprocessor instantiationawarebeanpostprocessor smartinstantiationawarebeanpostprocessor destructionawarebeanpostprocessor
1
5,476
8,354,656,652
IssuesEvent
2018-10-02 13:53:04
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
telomerase difference
PomBase cell cycle and DNA processes community curation mini-project
I'm checking a community submission related to telomerase. I don't understand the difference between these 2 terms: GO:0003720 telomerase activity Catalysis of the reaction: deoxynucleoside triphosphate + DNA(n) = diphosphate + DNA(n+1). Catalyzes extension of the 3'- end of a DNA strand by one deoxynucleotide at a time using an internal RNA template that encodes the telomeric repeat sequence. telomerase RNA reverse transcriptase activity Catalysis of the extension of the 3' end of a DNA strand by one deoxynucleotide at a time. Cannot initiate a chain de novo; uses the RNA subunit of the telomerase enzyme complex as its template.
1.0
telomerase difference - I'm checking a community submission related to telomerase. I don't understand the difference between these 2 terms: GO:0003720 telomerase activity Catalysis of the reaction: deoxynucleoside triphosphate + DNA(n) = diphosphate + DNA(n+1). Catalyzes extension of the 3'- end of a DNA strand by one deoxynucleotide at a time using an internal RNA template that encodes the telomeric repeat sequence. telomerase RNA reverse transcriptase activity Catalysis of the extension of the 3' end of a DNA strand by one deoxynucleotide at a time. Cannot initiate a chain de novo; uses the RNA subunit of the telomerase enzyme complex as its template.
process
telomerase difference i m checking a community submission related to telomerase i don t understand the difference between these terms go telomerase activity catalysis of the reaction deoxynucleoside triphosphate dna n diphosphate dna n catalyzes extension of the end of a dna strand by one deoxynucleotide at a time using an internal rna template that encodes the telomeric repeat sequence telomerase rna reverse transcriptase activity catalysis of the extension of the end of a dna strand by one deoxynucleotide at a time cannot initiate a chain de novo uses the rna subunit of the telomerase enzyme complex as its template
1
19,606
25,960,118,508
IssuesEvent
2022-12-18 19:48:15
streamnative/pulsar-spark
https://api.github.com/repos/streamnative/pulsar-spark
closed
[BUG] java.lang.RuntimeException: Failed to create schema for topic
type/bug compute/data-processing
When I try to call dataset. Write(), output the data to pulsar, and throw the error that failed to create the schema Environment: Pulsar-2.4.1 Spark-2.4.4 ``` java.lang.RuntimeException: Failed to create schema for persistent://public/default/als_statistics_tocheck at org.apache.spark.sql.pulsar.SchemaUtils$.uploadPulsarSchema(SchemaUtils.scala:104) at org.apache.spark.sql.pulsar.PulsarRowWriter.singleProducer$lzycompute(PulsarWriteTask.scala:140) at org.apache.spark.sql.pulsar.PulsarRowWriter.singleProducer(PulsarWriteTask.scala:138) at org.apache.spark.sql.pulsar.PulsarRowWriter.producerFlush(PulsarWriteTask.scala:210) ``` ``` RecordSchemaBuilder schemaBuilder = SchemaBuilder.record("topLevelRecord"); schemaBuilder.field("ip").type(SchemaType.STRING); schemaBuilder.field("port").type(SchemaType.INT32); schemaBuilder.field("url_id").type(SchemaType.STRING); schemaBuilder.field("response_rate").type(SchemaType.DOUBLE); schemaBuilder.field("success_rate").type(SchemaType.DOUBLE); schemaBuilder.field("average_response_time").type(SchemaType.DOUBLE); schemaBuilder.field("average_network_time").type(SchemaType.DOUBLE); schemaBuilder.field("start_time").type(SchemaType.TIMESTAMP); schemaBuilder.field("end_time").type(SchemaType.TIMESTAMP); SchemaInfo statistics2checkSchemaInfo = schemaBuilder.build(SchemaType.AVRO); waitToCheckDataSet.write() //.mode("append") .format("pulsar") .option("service.url", serviceUrl) .option("admin.url", adminUrl) .option("topic", statistics2CheckTopic) .option("pulsar.producer.sendTimeoutMs","60000") //.option("avroSchema",statistics2checkSchemaInfo.getSchemaDefinition()) //.option("recordName","CheckDataSet") //.option("recordNamespace","com.some.domain") .save(); try { admin.schemas().getSchemaInfo(statistics2CheckTopic); } catch (PulsarAdminException e) { if (404 == e.getStatusCode()) { admin.schemas().createSchema(statistics2CheckTopic,statistics2checkSchemaInfo); } } spark.readStream() .format("pulsar") .option("service.url", serviceUrl) .option("admin.url", adminUrl) .option("topic", statistics2CheckTopic) .option("startingOffsets", "earliest") .load() .withWatermark("__eventTime", "1 minute") .writeStream().queryName("WaitToCheckDataSet") .outputMode("append") .trigger(Trigger.ProcessingTime("1 minute")) .foreachBatch((dataset,batchId) -> { System.out.println("------WaitToCheckDataSet-------"); dataset.show(false); }).start(); ``` In addition,how can support options "avroSchema,recordName,recordNamespace" The name of the schema has always been named "topLevelRecord".Except : com.some.domain. ``` [http://spark.apache.org/docs/latest/sql-data-sources-avro.html](url) { "type": "record", "name": "topLevelRecord", "fields": [ { "name": "ip", "type": [ "string", "null" ] }, { "name": "port", "type": [ "int", "null" ] }, { "name": "url_id", "type": "string" }, { "name": "response_rate", "type": [ "double", "null" ] }, { "name": "success_rate", "type": [ "double", "null" ] }, { "name": "average_response_time", "type": "double" }, { "name": "average_network_time", "type": "double" }, { "name": "start_time", "type": [ { "type": "long", "logicalType": "timestamp-micros" }, "null" ] }, { "name": "end_time", "type": [ { "type": "long", "logicalType": "timestamp-micros" }, "null" ] }, { "name": "type", "type": "int" } ] } ```
1.0
[BUG] java.lang.RuntimeException: Failed to create schema for topic - When I try to call dataset. Write(), output the data to pulsar, and throw the error that failed to create the schema Environment: Pulsar-2.4.1 Spark-2.4.4 ``` java.lang.RuntimeException: Failed to create schema for persistent://public/default/als_statistics_tocheck at org.apache.spark.sql.pulsar.SchemaUtils$.uploadPulsarSchema(SchemaUtils.scala:104) at org.apache.spark.sql.pulsar.PulsarRowWriter.singleProducer$lzycompute(PulsarWriteTask.scala:140) at org.apache.spark.sql.pulsar.PulsarRowWriter.singleProducer(PulsarWriteTask.scala:138) at org.apache.spark.sql.pulsar.PulsarRowWriter.producerFlush(PulsarWriteTask.scala:210) ``` ``` RecordSchemaBuilder schemaBuilder = SchemaBuilder.record("topLevelRecord"); schemaBuilder.field("ip").type(SchemaType.STRING); schemaBuilder.field("port").type(SchemaType.INT32); schemaBuilder.field("url_id").type(SchemaType.STRING); schemaBuilder.field("response_rate").type(SchemaType.DOUBLE); schemaBuilder.field("success_rate").type(SchemaType.DOUBLE); schemaBuilder.field("average_response_time").type(SchemaType.DOUBLE); schemaBuilder.field("average_network_time").type(SchemaType.DOUBLE); schemaBuilder.field("start_time").type(SchemaType.TIMESTAMP); schemaBuilder.field("end_time").type(SchemaType.TIMESTAMP); SchemaInfo statistics2checkSchemaInfo = schemaBuilder.build(SchemaType.AVRO); waitToCheckDataSet.write() //.mode("append") .format("pulsar") .option("service.url", serviceUrl) .option("admin.url", adminUrl) .option("topic", statistics2CheckTopic) .option("pulsar.producer.sendTimeoutMs","60000") //.option("avroSchema",statistics2checkSchemaInfo.getSchemaDefinition()) //.option("recordName","CheckDataSet") //.option("recordNamespace","com.some.domain") .save(); try { admin.schemas().getSchemaInfo(statistics2CheckTopic); } catch (PulsarAdminException e) { if (404 == e.getStatusCode()) { admin.schemas().createSchema(statistics2CheckTopic,statistics2checkSchemaInfo); } } spark.readStream() .format("pulsar") .option("service.url", serviceUrl) .option("admin.url", adminUrl) .option("topic", statistics2CheckTopic) .option("startingOffsets", "earliest") .load() .withWatermark("__eventTime", "1 minute") .writeStream().queryName("WaitToCheckDataSet") .outputMode("append") .trigger(Trigger.ProcessingTime("1 minute")) .foreachBatch((dataset,batchId) -> { System.out.println("------WaitToCheckDataSet-------"); dataset.show(false); }).start(); ``` In addition,how can support options "avroSchema,recordName,recordNamespace" The name of the schema has always been named "topLevelRecord".Except : com.some.domain. ``` [http://spark.apache.org/docs/latest/sql-data-sources-avro.html](url) { "type": "record", "name": "topLevelRecord", "fields": [ { "name": "ip", "type": [ "string", "null" ] }, { "name": "port", "type": [ "int", "null" ] }, { "name": "url_id", "type": "string" }, { "name": "response_rate", "type": [ "double", "null" ] }, { "name": "success_rate", "type": [ "double", "null" ] }, { "name": "average_response_time", "type": "double" }, { "name": "average_network_time", "type": "double" }, { "name": "start_time", "type": [ { "type": "long", "logicalType": "timestamp-micros" }, "null" ] }, { "name": "end_time", "type": [ { "type": "long", "logicalType": "timestamp-micros" }, "null" ] }, { "name": "type", "type": "int" } ] } ```
process
java lang runtimeexception failed to create schema for topic when i try to call dataset write output the data to pulsar and throw the error that failed to create the schema environment pulsar spark java lang runtimeexception failed to create schema for persistent public default als statistics tocheck at org apache spark sql pulsar schemautils uploadpulsarschema schemautils scala at org apache spark sql pulsar pulsarrowwriter singleproducer lzycompute pulsarwritetask scala at org apache spark sql pulsar pulsarrowwriter singleproducer pulsarwritetask scala at org apache spark sql pulsar pulsarrowwriter producerflush pulsarwritetask scala recordschemabuilder schemabuilder schemabuilder record toplevelrecord schemabuilder field ip type schematype string schemabuilder field port type schematype schemabuilder field url id type schematype string schemabuilder field response rate type schematype double schemabuilder field success rate type schematype double schemabuilder field average response time type schematype double schemabuilder field average network time type schematype double schemabuilder field start time type schematype timestamp schemabuilder field end time type schematype timestamp schemainfo schemabuilder build schematype avro waittocheckdataset write mode append format pulsar option service url serviceurl option admin url adminurl option topic option pulsar producer sendtimeoutms option avroschema getschemadefinition option recordname checkdataset option recordnamespace com some domain save try admin schemas getschemainfo catch pulsaradminexception e if e getstatuscode admin schemas createschema spark readstream format pulsar option service url serviceurl option admin url adminurl option topic option startingoffsets earliest load withwatermark eventtime minute writestream queryname waittocheckdataset outputmode append trigger trigger processingtime minute foreachbatch dataset batchid system out println waittocheckdataset dataset show false start in addition how can support options avroschema recordname recordnamespace the name of the schema has always been named toplevelrecord except com some domain url type record name toplevelrecord fields name ip type string null name port type int null name url id type string name response rate type double null name success rate type double null name average response time type double name average network time type double name start time type type long logicaltype timestamp micros null name end time type type long logicaltype timestamp micros null name type type int
1
84,137
15,720,847,913
IssuesEvent
2021-03-29 01:24:06
wallanpsantos/microsservicos-java
https://api.github.com/repos/wallanpsantos/microsservicos-java
opened
CVE-2021-21349 (High) detected in xstream-1.4.13.jar
security vulnerability
## CVE-2021-21349 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.13.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: microsservicos-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.13/xstream-1.4.13.jar</p> <p> Dependency Hierarchy: - spring-cloud-starter-netflix-eureka-client-2.2.6.RELEASE.jar (Root Library) - :x: **xstream-1.4.13.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21349>CVE-2021-21349</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv">https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-21349 (High) detected in xstream-1.4.13.jar - ## CVE-2021-21349 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.13.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: microsservicos-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.13/xstream-1.4.13.jar</p> <p> Dependency Hierarchy: - spring-cloud-starter-netflix-eureka-client-2.2.6.RELEASE.jar (Root Library) - :x: **xstream-1.4.13.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21349>CVE-2021-21349</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv">https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in xstream jar cve high severity vulnerability vulnerable library xstream jar library home page a href path to dependency file microsservicos java pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy spring cloud starter netflix eureka client release jar root library x xstream jar vulnerable library found in base branch main vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream step up your open source security game with whitesource
0
18,039
5,557,347,378
IssuesEvent
2017-03-24 11:47:01
akvo/akvo-flow
https://api.github.com/repos/akvo/akvo-flow
opened
`Next` and `Previous` links are strangely aligned
Legacy clean-up & Code Refactoring
Under the inspect data tab <img width="271" alt="screen shot 2017-03-24 at 12 46 22" src="https://cloud.githubusercontent.com/assets/533336/24292984/e36d329e-108f-11e7-9f16-403e404eb538.png">
1.0
`Next` and `Previous` links are strangely aligned - Under the inspect data tab <img width="271" alt="screen shot 2017-03-24 at 12 46 22" src="https://cloud.githubusercontent.com/assets/533336/24292984/e36d329e-108f-11e7-9f16-403e404eb538.png">
non_process
next and previous links are strangely aligned under the inspect data tab img width alt screen shot at src
0
64,602
3,213,214,696
IssuesEvent
2015-10-06 18:52:06
cs2103aug2015-w15-2j/main
https://api.github.com/repos/cs2103aug2015-w15-2j/main
closed
A user can view events and tasks
priority.high type.story
so that the user can know what event is coming soon and what task is due soon.
1.0
A user can view events and tasks - so that the user can know what event is coming soon and what task is due soon.
non_process
a user can view events and tasks so that the user can know what event is coming soon and what task is due soon
0
13,828
16,592,340,356
IssuesEvent
2021-06-01 09:13:58
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Dropping Node v10 support (LTS reached EndOfLife)
process/candidate team/client tech/typescript topic: node version topic: prisma-client
For Prisma Client / Prisma CLI we need to drop Node.js v10 (now old LTS) Its End Of Life was on April 30 (for Security Support) and Active Support was dropped last year in May 2020. Happy to do it/pair on this since I know where changes are needed. The outcome will be that we support Node >=12.2.X and >=14.X (v16 is supported but not LTS yet, not recommended for production) See current version https://www.prisma.io/docs/reference/system-requirements ## Additional context https://prisma-company.slack.com/archives/C1FPU5FPT/p1621436107007300
1.0
Dropping Node v10 support (LTS reached EndOfLife) - For Prisma Client / Prisma CLI we need to drop Node.js v10 (now old LTS) Its End Of Life was on April 30 (for Security Support) and Active Support was dropped last year in May 2020. Happy to do it/pair on this since I know where changes are needed. The outcome will be that we support Node >=12.2.X and >=14.X (v16 is supported but not LTS yet, not recommended for production) See current version https://www.prisma.io/docs/reference/system-requirements ## Additional context https://prisma-company.slack.com/archives/C1FPU5FPT/p1621436107007300
process
dropping node support lts reached endoflife for prisma client prisma cli we need to drop node js now old lts its end of life was on april for security support and active support was dropped last year in may happy to do it pair on this since i know where changes are needed the outcome will be that we support node x and x is supported but not lts yet not recommended for production see current version additional context
1
13,672
16,419,346,646
IssuesEvent
2021-05-19 10:37:37
Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
https://api.github.com/repos/Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
closed
Benutzer soll sich registrieren können
backend frontend javascript register process
# Szenario: Benutzer soll sich registrieren können - **Gegeben** Der Benutzer ist auf der Startseite angelangt - **Wenn** sich der Benutzer registrieren möchte - **Dann** klickt er auf den Login Tab in der Navigation - **Und** es öffnet sich eine Registrierungsseite - **Und** der Nutzer registriert sich mit email und passwort ----- __Als__ Benutzer, __möchte ich__ mich registrieren können __damit__ ich mich einloggen kann. __Szenatio 1__: Benutzer kann sich registieren. __Szenario 2__: Registrierung schlägt fehl.
1.0
Benutzer soll sich registrieren können - # Szenario: Benutzer soll sich registrieren können - **Gegeben** Der Benutzer ist auf der Startseite angelangt - **Wenn** sich der Benutzer registrieren möchte - **Dann** klickt er auf den Login Tab in der Navigation - **Und** es öffnet sich eine Registrierungsseite - **Und** der Nutzer registriert sich mit email und passwort ----- __Als__ Benutzer, __möchte ich__ mich registrieren können __damit__ ich mich einloggen kann. __Szenatio 1__: Benutzer kann sich registieren. __Szenario 2__: Registrierung schlägt fehl.
process
benutzer soll sich registrieren können szenario benutzer soll sich registrieren können gegeben der benutzer ist auf der startseite angelangt wenn sich der benutzer registrieren möchte dann klickt er auf den login tab in der navigation und es öffnet sich eine registrierungsseite und der nutzer registriert sich mit email und passwort als benutzer möchte ich mich registrieren können damit ich mich einloggen kann szenatio benutzer kann sich registieren szenario registrierung schlägt fehl
1
21,284
28,479,167,507
IssuesEvent
2023-04-18 00:07:05
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Build multi-platform snapshot images
enhancement process
### Problem Currently snapshot images are being created and deployed in the registry, however they are build only for `amd64` platform. This causes inconvenience for people working on `arm64`, because those images are not working correctly, so they have to be build manually. ### Solution Enhance current release setup, to support cross-build. ### Alternatives _No response_
1.0
Build multi-platform snapshot images - ### Problem Currently snapshot images are being created and deployed in the registry, however they are build only for `amd64` platform. This causes inconvenience for people working on `arm64`, because those images are not working correctly, so they have to be build manually. ### Solution Enhance current release setup, to support cross-build. ### Alternatives _No response_
process
build multi platform snapshot images problem currently snapshot images are being created and deployed in the registry however they are build only for platform this causes inconvenience for people working on because those images are not working correctly so they have to be build manually solution enhance current release setup to support cross build alternatives no response
1
31,006
25,248,394,957
IssuesEvent
2022-11-15 12:56:42
aimclub/FEDOT
https://api.github.com/repos/aimclub/FEDOT
closed
Make logging output not so scary for end-user
enhancement infrastructure
During Graph verification (`GraphVerifier`) & Pipeline evaluation (`PipelineObjectiveEvaluate` with `EvaluationDispatcher`) usually happen a lot of errors. By default these errors are output to the logger and to the console. They bloat the console and create the impression that something goes terribly wrong. Fedot shouldn't create such impression and shouldn't by default output to the user information that's unhelpful, I propose that - console and log file must have different and configurable logging levels - by default console should output only informative messages about optimization process - all messages should by default go to the log file - in tests console output must output all verification & evaluation errors Regarding message log levels: - critical evaluation errors (which could lead to the failure of optimisation) should be errors - noncritical pipeline eval errors (from which optimiser recovers) should be warnings or debug-level messages (depending on severity) - verification errors seem to be expected errors and possibly should be debug-level messages
1.0
Make logging output not so scary for end-user - During Graph verification (`GraphVerifier`) & Pipeline evaluation (`PipelineObjectiveEvaluate` with `EvaluationDispatcher`) usually happen a lot of errors. By default these errors are output to the logger and to the console. They bloat the console and create the impression that something goes terribly wrong. Fedot shouldn't create such impression and shouldn't by default output to the user information that's unhelpful, I propose that - console and log file must have different and configurable logging levels - by default console should output only informative messages about optimization process - all messages should by default go to the log file - in tests console output must output all verification & evaluation errors Regarding message log levels: - critical evaluation errors (which could lead to the failure of optimisation) should be errors - noncritical pipeline eval errors (from which optimiser recovers) should be warnings or debug-level messages (depending on severity) - verification errors seem to be expected errors and possibly should be debug-level messages
non_process
make logging output not so scary for end user during graph verification graphverifier pipeline evaluation pipelineobjectiveevaluate with evaluationdispatcher usually happen a lot of errors by default these errors are output to the logger and to the console they bloat the console and create the impression that something goes terribly wrong fedot shouldn t create such impression and shouldn t by default output to the user information that s unhelpful i propose that console and log file must have different and configurable logging levels by default console should output only informative messages about optimization process all messages should by default go to the log file in tests console output must output all verification evaluation errors regarding message log levels critical evaluation errors which could lead to the failure of optimisation should be errors noncritical pipeline eval errors from which optimiser recovers should be warnings or debug level messages depending on severity verification errors seem to be expected errors and possibly should be debug level messages
0
133,722
29,507,101,009
IssuesEvent
2023-06-03 12:56:05
EddieHubCommunity/LinkFree
https://api.github.com/repos/EddieHubCommunity/LinkFree
closed
[OTHER] Remove single use mode
💻 aspect: code 🏁 status: ready for dev 🔢 points: 3
### What would you like to share? This feature is not used and we are keen for people to connect on the hosted platform REMOVE - layout - config - docs anything else? ### Additional information _No response_
1.0
[OTHER] Remove single use mode - ### What would you like to share? This feature is not used and we are keen for people to connect on the hosted platform REMOVE - layout - config - docs anything else? ### Additional information _No response_
non_process
remove single use mode what would you like to share this feature is not used and we are keen for people to connect on the hosted platform remove layout config docs anything else additional information no response
0
432,053
30,265,555,040
IssuesEvent
2023-07-07 11:34:17
WebXDAO/WebXGuild
https://api.github.com/repos/WebXDAO/WebXGuild
closed
[OTHER] Add Supabase setup guide into Readme
documentation good first issue
### What would you like to share? # Description We should update the documentation to add a setup guide for supabase. We will not share the main supabase credentials, contributors will need to create theirs. ## Details 1. We need to create a `schema.sql` file to share to contributors, they will use this schema to create their Supabase instance. 2. Setup a `.env.exemple` containing Supabase env keys 3. The guide should be detailed to simplify contributions. ### Additional information _No response_ ### Would you like to work on this issue? - [ ] Yes, I want to work on this issue! ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
[OTHER] Add Supabase setup guide into Readme - ### What would you like to share? # Description We should update the documentation to add a setup guide for supabase. We will not share the main supabase credentials, contributors will need to create theirs. ## Details 1. We need to create a `schema.sql` file to share to contributors, they will use this schema to create their Supabase instance. 2. Setup a `.env.exemple` containing Supabase env keys 3. The guide should be detailed to simplify contributions. ### Additional information _No response_ ### Would you like to work on this issue? - [ ] Yes, I want to work on this issue! ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_process
add supabase setup guide into readme what would you like to share description we should update the documentation to add a setup guide for supabase we will not share the main supabase credentials contributors will need to create theirs details we need to create a schema sql file to share to contributors they will use this schema to create their supabase instance setup a env exemple containing supabase env keys the guide should be detailed to simplify contributions additional information no response would you like to work on this issue yes i want to work on this issue code of conduct i agree to follow this project s code of conduct
0
58,803
14,484,859,357
IssuesEvent
2020-12-10 16:52:22
kubevirt/kubevirt
https://api.github.com/repos/kubevirt/kubevirt
closed
[Flaky CI] IOThreads IOThreads Policies [It] [test_id:4025]Should place io and emulator threads on the same pcpu with auto ioThreadsPolicy
kind/bug triage/build-watcher
**Is this a BUG REPORT or FEATURE REQUEST?**: /kind bug /triage build-watcher **What happened**: IOThreads IOThreads Policies [It] [test_id:4025]Should place io and emulator threads on the same pcpu with auto ioThreadsPolicy https://prow.apps.ovirt.org/view/gcs/kubevirt-prow/pr-logs/pull/kubevirt_kubevirt/4100/pull-kubevirt-e2e-k8s-1.16/1308357893728243712 ``` 13:15:23: IOThreads 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/vmi_iothreads_test.go:42 13:15:23: IOThreads Policies 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/vmi_iothreads_test.go:57 13:15:23: [test_id:4025]Should place io and emulator threads on the same pcpu with auto ioThreadsPolicy [It] 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/vmi_iothreads_test.go:229 13:15:23: 13:15:23: Timed out after 360.000s. 13:15:23: Timed out waiting for VMI testvmimvghllg4dhk55hzr65rjfh7xzcvtncp95vblgfvfzmgcnpwj to enter [Running] phase(s) 13:15:23: Expected 13:15:23: <v1.VirtualMachineInstancePhase>: Scheduling 13:15:23: to be an element of 13:15:23: <[]interface {} | len:1, cap:1>: [["Running"]] 13:15:23: 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/utils.go:2659 ```
1.0
[Flaky CI] IOThreads IOThreads Policies [It] [test_id:4025]Should place io and emulator threads on the same pcpu with auto ioThreadsPolicy - **Is this a BUG REPORT or FEATURE REQUEST?**: /kind bug /triage build-watcher **What happened**: IOThreads IOThreads Policies [It] [test_id:4025]Should place io and emulator threads on the same pcpu with auto ioThreadsPolicy https://prow.apps.ovirt.org/view/gcs/kubevirt-prow/pr-logs/pull/kubevirt_kubevirt/4100/pull-kubevirt-e2e-k8s-1.16/1308357893728243712 ``` 13:15:23: IOThreads 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/vmi_iothreads_test.go:42 13:15:23: IOThreads Policies 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/vmi_iothreads_test.go:57 13:15:23: [test_id:4025]Should place io and emulator threads on the same pcpu with auto ioThreadsPolicy [It] 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/vmi_iothreads_test.go:229 13:15:23: 13:15:23: Timed out after 360.000s. 13:15:23: Timed out waiting for VMI testvmimvghllg4dhk55hzr65rjfh7xzcvtncp95vblgfvfzmgcnpwj to enter [Running] phase(s) 13:15:23: Expected 13:15:23: <v1.VirtualMachineInstancePhase>: Scheduling 13:15:23: to be an element of 13:15:23: <[]interface {} | len:1, cap:1>: [["Running"]] 13:15:23: 13:15:23: /root/go/src/kubevirt.io/kubevirt/tests/utils.go:2659 ```
non_process
iothreads iothreads policies should place io and emulator threads on the same pcpu with auto iothreadspolicy is this a bug report or feature request kind bug triage build watcher what happened iothreads iothreads policies should place io and emulator threads on the same pcpu with auto iothreadspolicy iothreads root go src kubevirt io kubevirt tests vmi iothreads test go iothreads policies root go src kubevirt io kubevirt tests vmi iothreads test go should place io and emulator threads on the same pcpu with auto iothreadspolicy root go src kubevirt io kubevirt tests vmi iothreads test go timed out after timed out waiting for vmi to enter phase s expected scheduling to be an element of root go src kubevirt io kubevirt tests utils go
0
19,407
25,547,536,049
IssuesEvent
2022-11-29 20:12:34
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Add dependency check plugin to Gradle
enhancement process
### Problem To reach parity with Maven build we need to support scanning dependencies for security vulnerabilities. ### Solution * Add `org.owasp:dependency-check-gradle` plugin and configure it similarly as its Maven counterpart * Auto download Go and NPM before dependency check ### Alternatives _No response_
1.0
Add dependency check plugin to Gradle - ### Problem To reach parity with Maven build we need to support scanning dependencies for security vulnerabilities. ### Solution * Add `org.owasp:dependency-check-gradle` plugin and configure it similarly as its Maven counterpart * Auto download Go and NPM before dependency check ### Alternatives _No response_
process
add dependency check plugin to gradle problem to reach parity with maven build we need to support scanning dependencies for security vulnerabilities solution add org owasp dependency check gradle plugin and configure it similarly as its maven counterpart auto download go and npm before dependency check alternatives no response
1
143,134
5,510,755,922
IssuesEvent
2017-03-17 01:17:02
kamal1978/LTFHC
https://api.github.com/repos/kamal1978/LTFHC
closed
Register view not working
app bug high priority
Was originally able to view register. Not sure what changed that the table no longer exists (maybe because I cleared data?) ![screenshot_2017-03-15-21-24-21](https://cloud.githubusercontent.com/assets/9537700/23981700/cd4ee65e-09c5-11e7-9bf7-d49f902f0b2d.png)
1.0
Register view not working - Was originally able to view register. Not sure what changed that the table no longer exists (maybe because I cleared data?) ![screenshot_2017-03-15-21-24-21](https://cloud.githubusercontent.com/assets/9537700/23981700/cd4ee65e-09c5-11e7-9bf7-d49f902f0b2d.png)
non_process
register view not working was originally able to view register not sure what changed that the table no longer exists maybe because i cleared data
0
268,570
28,566,234,503
IssuesEvent
2023-04-21 02:31:25
Thanraj/packages_apps_Settings
https://api.github.com/repos/Thanraj/packages_apps_Settings
closed
CVE-2022-20533 (Low) detected in Settingsandroid-10.0.0_r44 - autoclosed
Mend: dependency security vulnerability
## CVE-2022-20533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r44</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p> <p>Found in HEAD commit: <a href="https://github.com/Thanraj/packages_apps_Settings/commit/cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa">cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/wifi/slice/WifiSlice.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In getSlice of WifiSlice.java, there is a possible way to connect a new WiFi network from the guest mode due to a missing permission check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-232798363 <p>Publish Date: 2022-12-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-20533>CVE-2022-20533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db">https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db</a></p> <p>Release Date: 2022-12-16</p> <p>Fix Resolution: android-13.0.0_r16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-20533 (Low) detected in Settingsandroid-10.0.0_r44 - autoclosed - ## CVE-2022-20533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r44</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p> <p>Found in HEAD commit: <a href="https://github.com/Thanraj/packages_apps_Settings/commit/cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa">cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/wifi/slice/WifiSlice.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In getSlice of WifiSlice.java, there is a possible way to connect a new WiFi network from the guest mode due to a missing permission check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-232798363 <p>Publish Date: 2022-12-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-20533>CVE-2022-20533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db">https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db</a></p> <p>Release Date: 2022-12-16</p> <p>Fix Resolution: android-13.0.0_r16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve low detected in settingsandroid autoclosed cve low severity vulnerability vulnerable library settingsandroid library home page a href found in head commit a href found in base branch master vulnerable source files src com android settings wifi slice wifislice java vulnerability details in getslice of wifislice java there is a possible way to connect a new wifi network from the guest mode due to a missing permission check this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
0
9,653
2,615,164,626
IssuesEvent
2015-03-01 06:44:42
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Can't associate using aireplay-ng
auto-migrated Priority-Triage Type-Defect
``` Hi I used aireplay to associate AP, but got a deauthentication packets over and over again... i don't know this problem, pls help me, tks! No source MAC (-h) specified. Using the device MAC (XXXXX) 23:16:02 Waiting for beacon frame (BSSID: XXXXX) on channel 6 23:16:02 Sending Authentication Request (Open System) [ACK] 23:16:02 Authentication successful 23:16:02 Sending Association Request [ACK] 23:16:02 Got a deauthentication packet! (Waiting 3 seconds) 23:16:05 Sending Authentication Request (Open System) [ACK] 23:16:05 Authentication successful 23:16:05 Sending Association Request [ACK] 23:16:05 Got a deauthentication packet! (Waiting 5 seconds) 23:16:10 Sending Authentication Request (Open System) [ACK] 23:16:10 Authentication successful 23:16:10 Sending Association Request [ACK] 23:16:10 Got a deauthentication packet! (Waiting 7 seconds) and my "wash -i mon0" show empty table: BSSID Channel RSSI WPS Version WPS Locked ESSID ---------------------------------------------------------- 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) 1.4 1. What operating system are you using (Linux is the only supported OS)? Ubuntu 12.10 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -36 4. What is the manufacturer and model # of the device you are trying to crack? TL-WR340G 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b bssid -vv -A aireplay-ng -1 30 -a bssid -e "network_name" mon0 wash -i mon0 6. Please describe what you think the issue is. No idea 7. Paste the outputs from Reaver below. output of aireplay above. ``` Original issue reported on code.google.com by `quangmai...@gmail.com` on 22 Oct 2012 at 5:01
1.0
Can't associate using aireplay-ng - ``` Hi I used aireplay to associate AP, but got a deauthentication packets over and over again... i don't know this problem, pls help me, tks! No source MAC (-h) specified. Using the device MAC (XXXXX) 23:16:02 Waiting for beacon frame (BSSID: XXXXX) on channel 6 23:16:02 Sending Authentication Request (Open System) [ACK] 23:16:02 Authentication successful 23:16:02 Sending Association Request [ACK] 23:16:02 Got a deauthentication packet! (Waiting 3 seconds) 23:16:05 Sending Authentication Request (Open System) [ACK] 23:16:05 Authentication successful 23:16:05 Sending Association Request [ACK] 23:16:05 Got a deauthentication packet! (Waiting 5 seconds) 23:16:10 Sending Authentication Request (Open System) [ACK] 23:16:10 Authentication successful 23:16:10 Sending Association Request [ACK] 23:16:10 Got a deauthentication packet! (Waiting 7 seconds) and my "wash -i mon0" show empty table: BSSID Channel RSSI WPS Version WPS Locked ESSID ---------------------------------------------------------- 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) 1.4 1. What operating system are you using (Linux is the only supported OS)? Ubuntu 12.10 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -36 4. What is the manufacturer and model # of the device you are trying to crack? TL-WR340G 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b bssid -vv -A aireplay-ng -1 30 -a bssid -e "network_name" mon0 wash -i mon0 6. Please describe what you think the issue is. No idea 7. Paste the outputs from Reaver below. output of aireplay above. ``` Original issue reported on code.google.com by `quangmai...@gmail.com` on 22 Oct 2012 at 5:01
non_process
can t associate using aireplay ng hi i used aireplay to associate ap but got a deauthentication packets over and over again i don t know this problem pls help me tks no source mac h specified using the device mac xxxxx waiting for beacon frame bssid xxxxx on channel sending authentication request open system authentication successful sending association request got a deauthentication packet waiting seconds sending authentication request open system authentication successful sending association request got a deauthentication packet waiting seconds sending authentication request open system authentication successful sending association request got a deauthentication packet waiting seconds and my wash i show empty table bssid channel rssi wps version wps locked essid what version of reaver are you using only defects against the latest version will be considered what operating system are you using linux is the only supported os ubuntu is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack tl what is the entire command line string you are supplying to reaver reaver i b bssid vv a aireplay ng a bssid e network name wash i please describe what you think the issue is no idea paste the outputs from reaver below output of aireplay above original issue reported on code google com by quangmai gmail com on oct at
0
18,340
24,462,894,128
IssuesEvent
2022-10-07 12:46:22
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
Add automatic input serialization to `ProcessFunction`s
type/accepted feature priority/nice-to-have topic/engine topic/processes
This will allow users to pass normal Python types to calculation and work functions ```python @calcfunction def add(x, y): return x + y add(1, 2) ```
1.0
Add automatic input serialization to `ProcessFunction`s - This will allow users to pass normal Python types to calculation and work functions ```python @calcfunction def add(x, y): return x + y add(1, 2) ```
process
add automatic input serialization to processfunction s this will allow users to pass normal python types to calculation and work functions python calcfunction def add x y return x y add
1
7,066
10,219,422,583
IssuesEvent
2019-08-15 18:32:36
toggl/mobileapp
https://api.github.com/repos/toggl/mobileapp
closed
Create template to submit "new copy" translation requests
process
Once we start supporting multiple languages. Adding new texts/copies throughout the app will need a little bit more of preparation. Ideally, new copy issues should be opened and closed before the new copies are close to be released. This template should be something in between #4835 and #4837, but it should contain the list of languages that need translation and should only be closed after all of them are merged into develop (or a project branch).
1.0
Create template to submit "new copy" translation requests - Once we start supporting multiple languages. Adding new texts/copies throughout the app will need a little bit more of preparation. Ideally, new copy issues should be opened and closed before the new copies are close to be released. This template should be something in between #4835 and #4837, but it should contain the list of languages that need translation and should only be closed after all of them are merged into develop (or a project branch).
process
create template to submit new copy translation requests once we start supporting multiple languages adding new texts copies throughout the app will need a little bit more of preparation ideally new copy issues should be opened and closed before the new copies are close to be released this template should be something in between and but it should contain the list of languages that need translation and should only be closed after all of them are merged into develop or a project branch
1
263,106
8,273,563,493
IssuesEvent
2018-09-17 06:34:08
edenlabllc/ehealth.api
https://api.github.com/repos/edenlabllc/ehealth.api
opened
address validation 500 error POST /api/declaration_requests
kind/bug priority/low
``` "phoenix": { "request_id": "2fe4bb61-5815-48fb-9050-7f8184ffac65#125909", "message": "** (CaseClauseError) no case clause matching: {:error, {:empty_body, 408}}\n (core) lib/core/validators/addresses.ex:38: Core.Validators.Addresses.validate_addresses_values/2\n (core) lib/core/declaration_requests/declaration_requests.ex:192: Core.DeclarationRequests.create_offline/2\n (ehealth) lib/ehealth/web/controllers/declaration_request_controller.ex:34: EHealth.Web.DeclarationRequestController.create/2\n (ehealth) lib/ehealth/web/controllers/declaration_request_controller.ex:1: EHealth.Web.DeclarationRequestController.action/2\n (ehealth) lib/ehealth/web/controllers/declaration_request_controller.ex:1: EHealth.Web.DeclarationRequestController.phoenix_controller_pipeline/2\n (ehealth) lib/ehealth/endpoint.ex:1: EHealth.Web.Endpoint.instrument/4\n (phoenix) lib/phoenix/router.ex:278: Phoenix.Router.__call__/1\n (ehealth) lib/plug/error_handler.ex:64: EHealthWeb.Router.call/2\n", "log_type": "error" }, "@timestamp": "2018-09-17T06:19:15.441676347+00:00", "tag": "phoenix.error" }, ```
1.0
address validation 500 error POST /api/declaration_requests - ``` "phoenix": { "request_id": "2fe4bb61-5815-48fb-9050-7f8184ffac65#125909", "message": "** (CaseClauseError) no case clause matching: {:error, {:empty_body, 408}}\n (core) lib/core/validators/addresses.ex:38: Core.Validators.Addresses.validate_addresses_values/2\n (core) lib/core/declaration_requests/declaration_requests.ex:192: Core.DeclarationRequests.create_offline/2\n (ehealth) lib/ehealth/web/controllers/declaration_request_controller.ex:34: EHealth.Web.DeclarationRequestController.create/2\n (ehealth) lib/ehealth/web/controllers/declaration_request_controller.ex:1: EHealth.Web.DeclarationRequestController.action/2\n (ehealth) lib/ehealth/web/controllers/declaration_request_controller.ex:1: EHealth.Web.DeclarationRequestController.phoenix_controller_pipeline/2\n (ehealth) lib/ehealth/endpoint.ex:1: EHealth.Web.Endpoint.instrument/4\n (phoenix) lib/phoenix/router.ex:278: Phoenix.Router.__call__/1\n (ehealth) lib/plug/error_handler.ex:64: EHealthWeb.Router.call/2\n", "log_type": "error" }, "@timestamp": "2018-09-17T06:19:15.441676347+00:00", "tag": "phoenix.error" }, ```
non_process
address validation error post api declaration requests phoenix request id message caseclauseerror no case clause matching error empty body n core lib core validators addresses ex core validators addresses validate addresses values n core lib core declaration requests declaration requests ex core declarationrequests create offline n ehealth lib ehealth web controllers declaration request controller ex ehealth web declarationrequestcontroller create n ehealth lib ehealth web controllers declaration request controller ex ehealth web declarationrequestcontroller action n ehealth lib ehealth web controllers declaration request controller ex ehealth web declarationrequestcontroller phoenix controller pipeline n ehealth lib ehealth endpoint ex ehealth web endpoint instrument n phoenix lib phoenix router ex phoenix router call n ehealth lib plug error handler ex ehealthweb router call n log type error timestamp tag phoenix error
0
2,300
5,116,796,496
IssuesEvent
2017-01-07 08:40:16
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Setting a class for edited cell on validation failure
help wanted inprocess
I currently use the code below to set edited cell color on validation failure. Would it be possible to allow this classname to be passed as part of table options and let the table do this on cell edit validation failure. I use the remote mode for the bootstrap table. Or is there a better way to achieve this? code: validator(){ if (document.querySelector('.editor.edit-text')) { document.querySelector('.editor.edit-text').parentElement.className = 'danger'; } return false; } **ScreenShot:** <img width="311" alt="cellvalidationfailure" src="https://cloud.githubusercontent.com/assets/13610968/21486964/c6e924b2-cb83-11e6-86d6-41345b77265a.png">
1.0
Setting a class for edited cell on validation failure - I currently use the code below to set edited cell color on validation failure. Would it be possible to allow this classname to be passed as part of table options and let the table do this on cell edit validation failure. I use the remote mode for the bootstrap table. Or is there a better way to achieve this? code: validator(){ if (document.querySelector('.editor.edit-text')) { document.querySelector('.editor.edit-text').parentElement.className = 'danger'; } return false; } **ScreenShot:** <img width="311" alt="cellvalidationfailure" src="https://cloud.githubusercontent.com/assets/13610968/21486964/c6e924b2-cb83-11e6-86d6-41345b77265a.png">
process
setting a class for edited cell on validation failure i currently use the code below to set edited cell color on validation failure would it be possible to allow this classname to be passed as part of table options and let the table do this on cell edit validation failure i use the remote mode for the bootstrap table or is there a better way to achieve this code validator if document queryselector editor edit text document queryselector editor edit text parentelement classname danger return false screenshot img width alt cellvalidationfailure src
1
344,763
10,349,602,525
IssuesEvent
2019-09-04 23:08:54
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Jupyter CR should support substatus
area/jupyter kind/feature priority/p1
/kind feature Jupyter custom resource needs to support sub resource per our application requirements (kubeflow/community#285). Its possible it already supports subresource in which case we just need to verify it and then we can close this issue. [Docs](https://kubernetes.io/docs/tasks/access-kubernetes-api/custom-resources/custom-resource-definitions/#status-subresource)
1.0
Jupyter CR should support substatus - /kind feature Jupyter custom resource needs to support sub resource per our application requirements (kubeflow/community#285). Its possible it already supports subresource in which case we just need to verify it and then we can close this issue. [Docs](https://kubernetes.io/docs/tasks/access-kubernetes-api/custom-resources/custom-resource-definitions/#status-subresource)
non_process
jupyter cr should support substatus kind feature jupyter custom resource needs to support sub resource per our application requirements kubeflow community its possible it already supports subresource in which case we just need to verify it and then we can close this issue
0
159,353
20,048,363,990
IssuesEvent
2022-02-03 01:09:25
kapseliboi/token-wizard
https://api.github.com/repos/kapseliboi/token-wizard
opened
CVE-2021-23807 (High) detected in jsonpointer-4.0.1.tgz
security vulnerability
## CVE-2021-23807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonpointer-4.0.1.tgz</b></p></summary> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/jsonpointer/package.json</p> <p> Dependency Hierarchy: - eslint-3.19.0.tgz (Root Library) - is-my-json-valid-2.17.1.tgz - :x: **jsonpointer-4.0.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package jsonpointer before 5.0.0. A type confusion vulnerability can lead to a bypass of a previous Prototype Pollution fix when the pointer components are arrays. <p>Publish Date: 2021-11-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807>CVE-2021-23807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807</a></p> <p>Release Date: 2021-11-03</p> <p>Fix Resolution (jsonpointer): 5.0.0</p> <p>Direct dependency fix Resolution (eslint): 4.0.0-alpha.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23807 (High) detected in jsonpointer-4.0.1.tgz - ## CVE-2021-23807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonpointer-4.0.1.tgz</b></p></summary> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/jsonpointer/package.json</p> <p> Dependency Hierarchy: - eslint-3.19.0.tgz (Root Library) - is-my-json-valid-2.17.1.tgz - :x: **jsonpointer-4.0.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package jsonpointer before 5.0.0. A type confusion vulnerability can lead to a bypass of a previous Prototype Pollution fix when the pointer components are arrays. <p>Publish Date: 2021-11-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23807>CVE-2021-23807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23807</a></p> <p>Release Date: 2021-11-03</p> <p>Fix Resolution (jsonpointer): 5.0.0</p> <p>Direct dependency fix Resolution (eslint): 4.0.0-alpha.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jsonpointer tgz cve high severity vulnerability vulnerable library jsonpointer tgz simple json addressing library home page a href path to dependency file package json path to vulnerable library node modules jsonpointer package json dependency hierarchy eslint tgz root library is my json valid tgz x jsonpointer tgz vulnerable library found in base branch master vulnerability details this affects the package jsonpointer before a type confusion vulnerability can lead to a bypass of a previous prototype pollution fix when the pointer components are arrays publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jsonpointer direct dependency fix resolution eslint alpha step up your open source security game with whitesource
0
18,967
24,938,379,535
IssuesEvent
2022-10-31 16:50:02
oxidecomputer/hubris
https://api.github.com/repos/oxidecomputer/hubris
closed
host-sp-comms set_status_impl() not taking power domain into account
service processor gimlet
The `host-sp-comms` tasks understandably exists across power domains. In particular, it can exist in A2 as it is part of the host normal path for executing commands like a reboot and related. However, this puts us in a particularly gnarly power domain case where we can possibly cross power domains. Normally the UART is protected by a buffer for the actual data and ready to send style pins that are treated as push-pull. However, the interrupt pin that we're using is not protected by that buffer. Unfortunately, all of the UART pins should only be driven from the ACPI state S0 according to the docs, which we treat as A0. Because the task can operate across domains the current construct in `set_status_impl()` is liable to cause back driving. UART backriving has been a serious concern and we've proven issues on Ethanol-X. While we do have isolation between the main UART signals, we didn't for the SP->SP3 interrupt line because at the time use was less clear and we figured the SP could do the right thing semantically. There are two approaches to this to consider: 1. `set_status_impl()` needs to be aware of the power domain and if we are transitioning away from A0 at all (more specifically the sequencer would toggle the uart enable lines) then we need to disable output on the GPIO. 2. Alternatively, because there is a pull up resistor on this, we may be able to get away with a bit of cheating and treating this like an open-drain pin on the SP side. Technically this is a push-pull from the SP3 side, but depending on if the interrupt is edge or level triggered and the rise times of this, that would allow us to potentially pretend the power domain doesn't exist as if the line is driven to 3.3V the pull up will cover it if it's in A0 and will not if we're not in A0 as the pull is in that domain.
1.0
host-sp-comms set_status_impl() not taking power domain into account - The `host-sp-comms` tasks understandably exists across power domains. In particular, it can exist in A2 as it is part of the host normal path for executing commands like a reboot and related. However, this puts us in a particularly gnarly power domain case where we can possibly cross power domains. Normally the UART is protected by a buffer for the actual data and ready to send style pins that are treated as push-pull. However, the interrupt pin that we're using is not protected by that buffer. Unfortunately, all of the UART pins should only be driven from the ACPI state S0 according to the docs, which we treat as A0. Because the task can operate across domains the current construct in `set_status_impl()` is liable to cause back driving. UART backriving has been a serious concern and we've proven issues on Ethanol-X. While we do have isolation between the main UART signals, we didn't for the SP->SP3 interrupt line because at the time use was less clear and we figured the SP could do the right thing semantically. There are two approaches to this to consider: 1. `set_status_impl()` needs to be aware of the power domain and if we are transitioning away from A0 at all (more specifically the sequencer would toggle the uart enable lines) then we need to disable output on the GPIO. 2. Alternatively, because there is a pull up resistor on this, we may be able to get away with a bit of cheating and treating this like an open-drain pin on the SP side. Technically this is a push-pull from the SP3 side, but depending on if the interrupt is edge or level triggered and the rise times of this, that would allow us to potentially pretend the power domain doesn't exist as if the line is driven to 3.3V the pull up will cover it if it's in A0 and will not if we're not in A0 as the pull is in that domain.
process
host sp comms set status impl not taking power domain into account the host sp comms tasks understandably exists across power domains in particular it can exist in as it is part of the host normal path for executing commands like a reboot and related however this puts us in a particularly gnarly power domain case where we can possibly cross power domains normally the uart is protected by a buffer for the actual data and ready to send style pins that are treated as push pull however the interrupt pin that we re using is not protected by that buffer unfortunately all of the uart pins should only be driven from the acpi state according to the docs which we treat as because the task can operate across domains the current construct in set status impl is liable to cause back driving uart backriving has been a serious concern and we ve proven issues on ethanol x while we do have isolation between the main uart signals we didn t for the sp interrupt line because at the time use was less clear and we figured the sp could do the right thing semantically there are two approaches to this to consider set status impl needs to be aware of the power domain and if we are transitioning away from at all more specifically the sequencer would toggle the uart enable lines then we need to disable output on the gpio alternatively because there is a pull up resistor on this we may be able to get away with a bit of cheating and treating this like an open drain pin on the sp side technically this is a push pull from the side but depending on if the interrupt is edge or level triggered and the rise times of this that would allow us to potentially pretend the power domain doesn t exist as if the line is driven to the pull up will cover it if it s in and will not if we re not in as the pull is in that domain
1
40,345
20,782,742,679
IssuesEvent
2022-03-16 16:06:53
sqlalchemy/sqlalchemy
https://api.github.com/repos/sqlalchemy/sqlalchemy
opened
reduce _is_clone_of depth for memory savings
bug orm performance
everytime we clone an expression it keeps around `_is_clone_of` as an intermediary, so that we have chains of expressions and columns that are all wasteful. _is_clone_of and _cloned_set are not consulted in too many places and it's very likely we only need it maintained for the root item, so we can at least flatten these. ```py import gc from sqlalchemy import Column from sqlalchemy import ForeignKey from sqlalchemy import Integer from sqlalchemy import String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy.orm.mapper import configure_mappers from sqlalchemy.sql.expression import BinaryExpression Base = declarative_base() class A(Base): __tablename__ = 'a' id = Column(Integer, primary_key=True) data = Column(String) bs = relationship("B") class B(Base): __tablename__ = 'b' id = Column(Integer, primary_key=True) a_id = Column(ForeignKey("a.id")) data = Column(String) configure_mappers() bc = A.bs.property.strategy._simple_lazy_clause[0] gc.collect() be = [x for x in gc.get_objects() if isinstance(x, BinaryExpression)] #breakpoint() print("binaries: %s" % len(be)) print("binaries: %s" % be) print("binaries: %s" % [str(x) for x in be]) ```
True
reduce _is_clone_of depth for memory savings - everytime we clone an expression it keeps around `_is_clone_of` as an intermediary, so that we have chains of expressions and columns that are all wasteful. _is_clone_of and _cloned_set are not consulted in too many places and it's very likely we only need it maintained for the root item, so we can at least flatten these. ```py import gc from sqlalchemy import Column from sqlalchemy import ForeignKey from sqlalchemy import Integer from sqlalchemy import String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy.orm.mapper import configure_mappers from sqlalchemy.sql.expression import BinaryExpression Base = declarative_base() class A(Base): __tablename__ = 'a' id = Column(Integer, primary_key=True) data = Column(String) bs = relationship("B") class B(Base): __tablename__ = 'b' id = Column(Integer, primary_key=True) a_id = Column(ForeignKey("a.id")) data = Column(String) configure_mappers() bc = A.bs.property.strategy._simple_lazy_clause[0] gc.collect() be = [x for x in gc.get_objects() if isinstance(x, BinaryExpression)] #breakpoint() print("binaries: %s" % len(be)) print("binaries: %s" % be) print("binaries: %s" % [str(x) for x in be]) ```
non_process
reduce is clone of depth for memory savings everytime we clone an expression it keeps around is clone of as an intermediary so that we have chains of expressions and columns that are all wasteful is clone of and cloned set are not consulted in too many places and it s very likely we only need it maintained for the root item so we can at least flatten these py import gc from sqlalchemy import column from sqlalchemy import foreignkey from sqlalchemy import integer from sqlalchemy import string from sqlalchemy ext declarative import declarative base from sqlalchemy orm import relationship from sqlalchemy orm mapper import configure mappers from sqlalchemy sql expression import binaryexpression base declarative base class a base tablename a id column integer primary key true data column string bs relationship b class b base tablename b id column integer primary key true a id column foreignkey a id data column string configure mappers bc a bs property strategy simple lazy clause gc collect be breakpoint print binaries s len be print binaries s be print binaries s
0
18,255
24,335,641,678
IssuesEvent
2022-10-01 03:16:57
GoogleCloudPlatform/cloud-ops-sandbox
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
opened
Create test configurations to be used in e2e testing
priority: p1 type: process
Create initial stubs of the test configurations to be used in e2e testing: * `tests/test-app/monitoring.json` * `tests/test-app/alerting.json` * `tests/test-app/services.json` * `tests/test-app/endpoints.json` * `tests/test-app/recipes.json`
1.0
Create test configurations to be used in e2e testing - Create initial stubs of the test configurations to be used in e2e testing: * `tests/test-app/monitoring.json` * `tests/test-app/alerting.json` * `tests/test-app/services.json` * `tests/test-app/endpoints.json` * `tests/test-app/recipes.json`
process
create test configurations to be used in testing create initial stubs of the test configurations to be used in testing tests test app monitoring json tests test app alerting json tests test app services json tests test app endpoints json tests test app recipes json
1
13,996
16,768,493,766
IssuesEvent
2021-06-14 12:04:59
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Android] Text in Data sharing screen is not displaying as configured in Study builder
Android Bug P2 Process: Fixed Process: Tested QA Process: Tested dev
**Steps**:- 1. Configure the Data sharing in E-consent screen in the Study builder as shown below. 2. Login into android application 3. Navigate to the studies list 4. Click on study and try to enroll 5. Navigate to Data sharing screen and click on **Learn more** 6. Verify the text in Learn more screen **A/R**:- Text is displaying in a single line **E/R**:- Text should display as configured in the Study builder ![image](https://user-images.githubusercontent.com/60500517/119983826-4d922380-bfde-11eb-8e44-830ed294baf6.png) ![image](https://user-images.githubusercontent.com/60500517/119983859-5aaf1280-bfde-11eb-86d2-8be3699e28c8.png)
3.0
[Android] Text in Data sharing screen is not displaying as configured in Study builder - **Steps**:- 1. Configure the Data sharing in E-consent screen in the Study builder as shown below. 2. Login into android application 3. Navigate to the studies list 4. Click on study and try to enroll 5. Navigate to Data sharing screen and click on **Learn more** 6. Verify the text in Learn more screen **A/R**:- Text is displaying in a single line **E/R**:- Text should display as configured in the Study builder ![image](https://user-images.githubusercontent.com/60500517/119983826-4d922380-bfde-11eb-8e44-830ed294baf6.png) ![image](https://user-images.githubusercontent.com/60500517/119983859-5aaf1280-bfde-11eb-86d2-8be3699e28c8.png)
process
text in data sharing screen is not displaying as configured in study builder steps configure the data sharing in e consent screen in the study builder as shown below login into android application navigate to the studies list click on study and try to enroll navigate to data sharing screen and click on learn more verify the text in learn more screen a r text is displaying in a single line e r text should display as configured in the study builder
1
167,421
13,024,815,524
IssuesEvent
2020-07-27 12:32:41
googleapis/google-cloud-cpp
https://api.github.com/repos/googleapis/google-cloud-cpp
closed
implement testing `Status` matchers
priority: p3 testing type: feature request
Implement matchers that can check a `google::cloud::Status` in a single `EXPECT` statement. We shouldn't spend too much effort on this until we resolve whether to switch to `absl::Status` (#4375), but we can implement our matchers in ways that are similar to (more specifically, subsets of) absl matchers to help the transition if we do go that direction.
1.0
implement testing `Status` matchers - Implement matchers that can check a `google::cloud::Status` in a single `EXPECT` statement. We shouldn't spend too much effort on this until we resolve whether to switch to `absl::Status` (#4375), but we can implement our matchers in ways that are similar to (more specifically, subsets of) absl matchers to help the transition if we do go that direction.
non_process
implement testing status matchers implement matchers that can check a google cloud status in a single expect statement we shouldn t spend too much effort on this until we resolve whether to switch to absl status but we can implement our matchers in ways that are similar to more specifically subsets of absl matchers to help the transition if we do go that direction
0
126,207
12,288,346,276
IssuesEvent
2020-05-09 16:19:33
ISPPNightTurn/Clubby
https://api.github.com/repos/ISPPNightTurn/Clubby
closed
Aclaración sobre las características de Clubby
documentation
Redactar de forma clara las caracteristicas y funcionalidades de Clubby dividiendola en dos partes según cliente y usuario. Es necesario para llevar a cabo las campañas de marketing más eficientemente. También para dejar claro a todos los componentes del equipo que tiene nuestra aplicación en caso de que no lo sepan ahora que el desarrollo ya ha terminado.
1.0
Aclaración sobre las características de Clubby - Redactar de forma clara las caracteristicas y funcionalidades de Clubby dividiendola en dos partes según cliente y usuario. Es necesario para llevar a cabo las campañas de marketing más eficientemente. También para dejar claro a todos los componentes del equipo que tiene nuestra aplicación en caso de que no lo sepan ahora que el desarrollo ya ha terminado.
non_process
aclaración sobre las características de clubby redactar de forma clara las caracteristicas y funcionalidades de clubby dividiendola en dos partes según cliente y usuario es necesario para llevar a cabo las campañas de marketing más eficientemente también para dejar claro a todos los componentes del equipo que tiene nuestra aplicación en caso de que no lo sepan ahora que el desarrollo ya ha terminado
0
137,760
20,206,749,453
IssuesEvent
2022-02-11 21:24:29
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
#endif is seen as LeadingTrivia of namespace instead of TrailingTrivia of using statement
Area-Compilers Resolution-By Design untriaged
**Version Used**: **Steps to Reproduce**: 1. Using statement surrounded by `#if` preprocessor directive ``` using System; #if UNITY using Unity #endif namespace Namespace { public class ClassName { } } ``` 2. Analyze code in roslyn with source generator. 3. Use source generator to move `UsingStatementSyntax` inside of the `NamespaceDeclartionSyntax` ``` Console.WriteLine( namespaceSyntax.WithUsingStatements( CompilationUnitSyntax.Usings ); ``` **Expected Behavior**: ``` namespace Namespace { using System; #if UNITY using Unity #endif public class ClassName { } } ``` **Actual Behavior**: ``` #endif namespace Namespace { using System; #if UNITY using Unity public class ClassName { } } ``` Exploring the syntax tree shows the `#endif` trivia is the `LeadingTrivia` of the `NamespaceDeclarationSyntax` and not the `TrailingTrivia` of the `UsingStatementSyntax`. I think it was implemented this way to cover the edge case where there is no token in front of the `#if` directive by assuming there will always be a NamespaceDeclaration, UsingStatment or EndOfFile token after it. However this can lead to "unintended behavior" as the formatting becomes incorrect when it attaches the `#if` and `#endif` to different tokens and you try moving those tokens around.
1.0
#endif is seen as LeadingTrivia of namespace instead of TrailingTrivia of using statement - **Version Used**: **Steps to Reproduce**: 1. Using statement surrounded by `#if` preprocessor directive ``` using System; #if UNITY using Unity #endif namespace Namespace { public class ClassName { } } ``` 2. Analyze code in roslyn with source generator. 3. Use source generator to move `UsingStatementSyntax` inside of the `NamespaceDeclartionSyntax` ``` Console.WriteLine( namespaceSyntax.WithUsingStatements( CompilationUnitSyntax.Usings ); ``` **Expected Behavior**: ``` namespace Namespace { using System; #if UNITY using Unity #endif public class ClassName { } } ``` **Actual Behavior**: ``` #endif namespace Namespace { using System; #if UNITY using Unity public class ClassName { } } ``` Exploring the syntax tree shows the `#endif` trivia is the `LeadingTrivia` of the `NamespaceDeclarationSyntax` and not the `TrailingTrivia` of the `UsingStatementSyntax`. I think it was implemented this way to cover the edge case where there is no token in front of the `#if` directive by assuming there will always be a NamespaceDeclaration, UsingStatment or EndOfFile token after it. However this can lead to "unintended behavior" as the formatting becomes incorrect when it attaches the `#if` and `#endif` to different tokens and you try moving those tokens around.
non_process
endif is seen as leadingtrivia of namespace instead of trailingtrivia of using statement version used steps to reproduce using statement surrounded by if preprocessor directive using system if unity using unity endif namespace namespace public class classname analyze code in roslyn with source generator use source generator to move usingstatementsyntax inside of the namespacedeclartionsyntax console writeline namespacesyntax withusingstatements compilationunitsyntax usings expected behavior namespace namespace using system if unity using unity endif public class classname actual behavior endif namespace namespace using system if unity using unity public class classname exploring the syntax tree shows the endif trivia is the leadingtrivia of the namespacedeclarationsyntax and not the trailingtrivia of the usingstatementsyntax i think it was implemented this way to cover the edge case where there is no token in front of the if directive by assuming there will always be a namespacedeclaration usingstatment or endoffile token after it however this can lead to unintended behavior as the formatting becomes incorrect when it attaches the if and endif to different tokens and you try moving those tokens around
0
334,411
10,141,675,073
IssuesEvent
2019-08-03 16:21:51
SacredDuckwhale/Rarity
https://api.github.com/repos/SacredDuckwhale/Rarity
closed
Add supplies from New Pet Battle Dungeon: Stratholme
complexity: low module:database priority:high status:needs-confirmation type:task
Source: https://www.mmo-champion.com/content/8556-Stratholme-Pet-Battle-Dungeon-Preview-World-Quest-Bonus-Event-Deepwind-Dunk-Brawl Not sure when/if this will be added, but I'll have to look into it eventually. Just putting this here so I won't forget.
1.0
Add supplies from New Pet Battle Dungeon: Stratholme - Source: https://www.mmo-champion.com/content/8556-Stratholme-Pet-Battle-Dungeon-Preview-World-Quest-Bonus-Event-Deepwind-Dunk-Brawl Not sure when/if this will be added, but I'll have to look into it eventually. Just putting this here so I won't forget.
non_process
add supplies from new pet battle dungeon stratholme source not sure when if this will be added but i ll have to look into it eventually just putting this here so i won t forget
0
14,625
17,766,987,319
IssuesEvent
2021-08-30 08:48:55
googleapis/jsdoc-region-tag
https://api.github.com/repos/googleapis/jsdoc-region-tag
closed
Dependency Dashboard
type: process
This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/61) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/61) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull check this box to trigger a request for renovate to run again on this repository
1
15,244
19,182,515,014
IssuesEvent
2021-12-04 16:53:39
RobertCraigie/prisma-client-py
https://api.github.com/repos/RobertCraigie/prisma-client-py
closed
DateTime precision loss leads to records being unable to be found
bug/2-confirmed kind/bug process/candidate
<!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> Internally, Prisma truncates `DateTime` values to 3 decimal places, e.g. `2021-11-17 17:04:56.936198` -> `2021-11-17 17:04:56.936000` This can lead to a bug where records cannot be filtered by a `DateTime` equals value. This bug was not caught until now as it does not appear on SQLite but does on PostgreSQL. ## How to reproduce <!-- Steps to reproduce the behavior: 1. Go to '...' 2. Change '....' 3. Run '....' 4. See error --> ```py import asyncio from datetime import datetime from prisma import Client async def main() -> None: db = Client() await db.connect() date = datetime.utcnow() print(date) post = await db.post.create( data={ 'title': 'My first post', 'created_at': date, }, ) print(post.created_at) found = await db.post.find_first( where={ 'created_at': date, }, ) print(f'Post found? {found is not None}') await db.disconnect() if __name__ == '__main__': asyncio.get_event_loop().run_until_complete(main()) ``` SQLite output: ``` 2021-11-17 17:13:21.288983 2021-11-17 17:13:21.288000+00:00 Post found? True ``` PostgreSQL output: ``` 2021-11-17 17:14:15.055458 2021-11-17 17:14:15.055000+00:00 Post found? False ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> PostgreSQL output: ``` 2021-11-17 17:14:15.055458 2021-11-17 17:14:15.055000+00:00 Post found? True ``` ## Prisma information <!-- Your Prisma schema, Prisma Client Python queries, ... Do not include your database credentials when sharing your Prisma schema! --> ```prisma datasource client { // provider = "sqlite" // url = "file:dev.db" provider = "postgres" url = env("DATABASE_URL") } generator client { provider = "prisma-client-py" interface = "asyncio" } model Post { id String @id @default(cuid()) created_at DateTime @default(now()) updated_at DateTime @updatedAt title String published Boolean @default(false) desc String? } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> Mac OS - Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> PostgreSQL - Python version: <!--[Run `python -V` to see your Python version]--> Python 3.9.1 - Prisma version: <!--[Run `prisma py version` to see your Prisma version and paste it between the ´´´]--> ``` prisma : 3.4.0 prisma client python : 0.3.0 platform : darwin engines : 1c9fdaa9e2319b814822d6dbfd0a69e1fcc13a85 install path : .venv/lib/python3.9/site-packages/prisma installed extras : [] ```
1.0
DateTime precision loss leads to records being unable to be found - <!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> Internally, Prisma truncates `DateTime` values to 3 decimal places, e.g. `2021-11-17 17:04:56.936198` -> `2021-11-17 17:04:56.936000` This can lead to a bug where records cannot be filtered by a `DateTime` equals value. This bug was not caught until now as it does not appear on SQLite but does on PostgreSQL. ## How to reproduce <!-- Steps to reproduce the behavior: 1. Go to '...' 2. Change '....' 3. Run '....' 4. See error --> ```py import asyncio from datetime import datetime from prisma import Client async def main() -> None: db = Client() await db.connect() date = datetime.utcnow() print(date) post = await db.post.create( data={ 'title': 'My first post', 'created_at': date, }, ) print(post.created_at) found = await db.post.find_first( where={ 'created_at': date, }, ) print(f'Post found? {found is not None}') await db.disconnect() if __name__ == '__main__': asyncio.get_event_loop().run_until_complete(main()) ``` SQLite output: ``` 2021-11-17 17:13:21.288983 2021-11-17 17:13:21.288000+00:00 Post found? True ``` PostgreSQL output: ``` 2021-11-17 17:14:15.055458 2021-11-17 17:14:15.055000+00:00 Post found? False ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> PostgreSQL output: ``` 2021-11-17 17:14:15.055458 2021-11-17 17:14:15.055000+00:00 Post found? True ``` ## Prisma information <!-- Your Prisma schema, Prisma Client Python queries, ... Do not include your database credentials when sharing your Prisma schema! --> ```prisma datasource client { // provider = "sqlite" // url = "file:dev.db" provider = "postgres" url = env("DATABASE_URL") } generator client { provider = "prisma-client-py" interface = "asyncio" } model Post { id String @id @default(cuid()) created_at DateTime @default(now()) updated_at DateTime @updatedAt title String published Boolean @default(false) desc String? } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> Mac OS - Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> PostgreSQL - Python version: <!--[Run `python -V` to see your Python version]--> Python 3.9.1 - Prisma version: <!--[Run `prisma py version` to see your Prisma version and paste it between the ´´´]--> ``` prisma : 3.4.0 prisma client python : 0.3.0 platform : darwin engines : 1c9fdaa9e2319b814822d6dbfd0a69e1fcc13a85 install path : .venv/lib/python3.9/site-packages/prisma installed extras : [] ```
process
datetime precision loss leads to records being unable to be found thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description internally prisma truncates datetime values to decimal places e g this can lead to a bug where records cannot be filtered by a datetime equals value this bug was not caught until now as it does not appear on sqlite but does on postgresql how to reproduce steps to reproduce the behavior go to change run see error py import asyncio from datetime import datetime from prisma import client async def main none db client await db connect date datetime utcnow print date post await db post create data title my first post created at date print post created at found await db post find first where created at date print f post found found is not none await db disconnect if name main asyncio get event loop run until complete main sqlite output post found true postgresql output post found false expected behavior postgresql output post found true prisma information your prisma schema prisma client python queries do not include your database credentials when sharing your prisma schema prisma datasource client provider sqlite url file dev db provider postgres url env database url generator client provider prisma client py interface asyncio model post id string id default cuid created at datetime default now updated at datetime updatedat title string published boolean default false desc string environment setup os mac os database postgresql python version python prisma version prisma prisma client python platform darwin engines install path venv lib site packages prisma installed extras
1
75,257
25,724,982,710
IssuesEvent
2022-12-07 16:04:04
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Replying to root message in thread causes right panel to go blank
T-Defect A-Threads
### Steps to reproduce 1. Go to #publicencryptedroom:matrix.org and post a message 2. Click the thread button in the main timeline 3. In the right panel, click the reply button ### Outcome #### What did you expect? Can reply to message #### What happened instead? Root message vanishes from right panel and it looks empty ![Screenshot from 2022-12-07 16-03-55](https://user-images.githubusercontent.com/51663/206229458-b76696a4-1066-4ae8-9293-60dec04aef18.png) ### Operating system _No response_ ### Browser information Chromium 108.0.5359.94 (Official Build) Arch Linux (64-bit) ### URL for webapp develop.element.io ### Application version Element version: c473cb6-react-c3809d3afa7d-js-8d018f9c2d33 Olm version: 3.2.12 ### Homeserver matrix.org ### Will you send logs? Yes
1.0
Replying to root message in thread causes right panel to go blank - ### Steps to reproduce 1. Go to #publicencryptedroom:matrix.org and post a message 2. Click the thread button in the main timeline 3. In the right panel, click the reply button ### Outcome #### What did you expect? Can reply to message #### What happened instead? Root message vanishes from right panel and it looks empty ![Screenshot from 2022-12-07 16-03-55](https://user-images.githubusercontent.com/51663/206229458-b76696a4-1066-4ae8-9293-60dec04aef18.png) ### Operating system _No response_ ### Browser information Chromium 108.0.5359.94 (Official Build) Arch Linux (64-bit) ### URL for webapp develop.element.io ### Application version Element version: c473cb6-react-c3809d3afa7d-js-8d018f9c2d33 Olm version: 3.2.12 ### Homeserver matrix.org ### Will you send logs? Yes
non_process
replying to root message in thread causes right panel to go blank steps to reproduce go to publicencryptedroom matrix org and post a message click the thread button in the main timeline in the right panel click the reply button outcome what did you expect can reply to message what happened instead root message vanishes from right panel and it looks empty operating system no response browser information chromium official build arch linux bit url for webapp develop element io application version element version react js olm version homeserver matrix org will you send logs yes
0
247,051
20,955,951,751
IssuesEvent
2022-03-27 04:59:41
cricarba/isolucionStatus
https://api.github.com/repos/cricarba/isolucionStatus
closed
🛑 lamesaTest.isolucion.co is down
status lamesa-test-isolucion-co
In [`112f359`](https://github.com/cricarba/isolucionStatus/commit/112f359fbcac9dc24909430778be553325b92201 ), lamesaTest.isolucion.co (https://lamesaTest.isolucion.co) was **down**: - HTTP code: 0 - Response time: 0 ms
1.0
🛑 lamesaTest.isolucion.co is down - In [`112f359`](https://github.com/cricarba/isolucionStatus/commit/112f359fbcac9dc24909430778be553325b92201 ), lamesaTest.isolucion.co (https://lamesaTest.isolucion.co) was **down**: - HTTP code: 0 - Response time: 0 ms
non_process
🛑 lamesatest isolucion co is down in lamesatest isolucion co was down http code response time ms
0
10,122
13,044,162,270
IssuesEvent
2020-07-29 03:47:31
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `StringDurationTimeDiff` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `StringDurationTimeDiff` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `StringDurationTimeDiff` from TiDB - ## Description Port the scalar function `StringDurationTimeDiff` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function stringdurationtimediff from tidb description port the scalar function stringdurationtimediff from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
1
263,732
8,301,197,197
IssuesEvent
2018-09-21 10:32:34
bigbinary/acehelp
https://api.github.com/repos/bigbinary/acehelp
closed
When user click on Logout two request are being sent to the server
Priority Issue Working
> 2018-09-20T12:29:20.131147+00:00 heroku[router]: **at=info method=OPTIONS path="/users/sign_out"** host=staging.acehelp.com request_id=e612d4e2-39fb-47c3-bba2-984d4fcfcb8e fwd="123.201.36.71" dyno=web.1 connect=1ms service=3ms status=200 bytes=271 protocol=https 2018-09-20T12:29:20.375261+00:00 heroku[router]: **at=info method=GET path="/organizations/3ef4539f24e4a9fb6d20/articles" host=staging.acehelp.com request_id=c1164d74-09d0-4ec5-a920-8b0983c9a3b3** fwd="123.201.36.71" dyno=web.1 connect=1ms service=14ms status=302 bytes=1061 protocol=http 2018-09-20T12:29:20.363107+00:00 app[web.1]: I, [2018-09-20T12:29:20.362944 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Started GET "/organizations/3ef4539f24e4a9fb6d20/articles" for 123.201.36.71 at 2018-09-20 12:29:20 +0000 2018-09-20T12:29:20.364542+00:00 app[web.1]: I, [2018-09-20T12:29:20.364439 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] **Processing by ArticlesController#index as HTML 2018-09-20T12:29:20.364641+00:00 app[web.1]: I, [2018-09-20T12:29:20.364548 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Parameters: {"organization_api_key"=>"3ef4539f24e4a9fb6d20"}** 2018-09-20T12:29:20.373405+00:00 app[web.1]: I, [2018-09-20T12:29:20.373279 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Redirected to http://staging.acehelp.com/users/sign_in 2018-09-20T12:29:20.373575+00:00 app[web.1]: I, [2018-09-20T12:29:20.373496 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Filter chain halted as :ensure_user_is_logged_in rendered or redirected 2018-09-20T12:29:20.373813+00:00 app[web.1]: I, [2018-09-20T12:29:20.373731 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Completed 302 Found in 9ms (ActiveRecord: 0.0ms) 2018-09-20T12:29:20.848477+00:00 heroku[router]: at=info method=GET path="/users/sign_in" host=staging.acehelp.com request_id=6bf0c608-7522-4513-a0e7-a78ead98027b fwd="123.201.36.71" dyno=web.1 connect=0ms service=14ms status=200 bytes=1682 protocol=http << This is the log from staging instance. After the user clicks on logout two requests are coming to the server. @prathamesh-sonpatki @ritesh404 Please have a look at this. Ritesh: Can you help to figure out why might be this case. We are getting two requests at the server when we click on `Logout`.
1.0
When user click on Logout two request are being sent to the server - > 2018-09-20T12:29:20.131147+00:00 heroku[router]: **at=info method=OPTIONS path="/users/sign_out"** host=staging.acehelp.com request_id=e612d4e2-39fb-47c3-bba2-984d4fcfcb8e fwd="123.201.36.71" dyno=web.1 connect=1ms service=3ms status=200 bytes=271 protocol=https 2018-09-20T12:29:20.375261+00:00 heroku[router]: **at=info method=GET path="/organizations/3ef4539f24e4a9fb6d20/articles" host=staging.acehelp.com request_id=c1164d74-09d0-4ec5-a920-8b0983c9a3b3** fwd="123.201.36.71" dyno=web.1 connect=1ms service=14ms status=302 bytes=1061 protocol=http 2018-09-20T12:29:20.363107+00:00 app[web.1]: I, [2018-09-20T12:29:20.362944 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Started GET "/organizations/3ef4539f24e4a9fb6d20/articles" for 123.201.36.71 at 2018-09-20 12:29:20 +0000 2018-09-20T12:29:20.364542+00:00 app[web.1]: I, [2018-09-20T12:29:20.364439 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] **Processing by ArticlesController#index as HTML 2018-09-20T12:29:20.364641+00:00 app[web.1]: I, [2018-09-20T12:29:20.364548 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Parameters: {"organization_api_key"=>"3ef4539f24e4a9fb6d20"}** 2018-09-20T12:29:20.373405+00:00 app[web.1]: I, [2018-09-20T12:29:20.373279 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Redirected to http://staging.acehelp.com/users/sign_in 2018-09-20T12:29:20.373575+00:00 app[web.1]: I, [2018-09-20T12:29:20.373496 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Filter chain halted as :ensure_user_is_logged_in rendered or redirected 2018-09-20T12:29:20.373813+00:00 app[web.1]: I, [2018-09-20T12:29:20.373731 #4] INFO -- : [c1164d74-09d0-4ec5-a920-8b0983c9a3b3] Completed 302 Found in 9ms (ActiveRecord: 0.0ms) 2018-09-20T12:29:20.848477+00:00 heroku[router]: at=info method=GET path="/users/sign_in" host=staging.acehelp.com request_id=6bf0c608-7522-4513-a0e7-a78ead98027b fwd="123.201.36.71" dyno=web.1 connect=0ms service=14ms status=200 bytes=1682 protocol=http << This is the log from staging instance. After the user clicks on logout two requests are coming to the server. @prathamesh-sonpatki @ritesh404 Please have a look at this. Ritesh: Can you help to figure out why might be this case. We are getting two requests at the server when we click on `Logout`.
non_process
when user click on logout two request are being sent to the server heroku at info method options path users sign out host staging acehelp com request id fwd dyno web connect service status bytes protocol https heroku at info method get path organizations articles host staging acehelp com request id fwd dyno web connect service status bytes protocol http app i info started get organizations articles for at app i info processing by articlescontroller index as html app i info parameters organization api key app i info redirected to app i info filter chain halted as ensure user is logged in rendered or redirected app i info completed found in activerecord heroku at info method get path users sign in host staging acehelp com request id fwd dyno web connect service status bytes protocol http this is the log from staging instance after the user clicks on logout two requests are coming to the server prathamesh sonpatki please have a look at this ritesh can you help to figure out why might be this case we are getting two requests at the server when we click on logout
0
5,298
8,120,640,427
IssuesEvent
2018-08-16 04:01:46
google/codeworld
https://api.github.com/repos/google/codeworld
closed
Transform on canvas is spuriously lost
bug in process
The program at https://code.world/#PRheJxUqzHjsy1I59l3pFhA acts oddly. It should draw an approximate circle out of random triangles. Instead, at least on Chrome 62, part of the circle is moved to the top-left corner of the screen. It's almost as if the Context2D transform applied at the start of the drawing is forgotten partway through. ![image](https://user-images.githubusercontent.com/544744/33791176-e8cd31fc-dc3c-11e7-911a-25e72d8cf13d.png)
1.0
Transform on canvas is spuriously lost - The program at https://code.world/#PRheJxUqzHjsy1I59l3pFhA acts oddly. It should draw an approximate circle out of random triangles. Instead, at least on Chrome 62, part of the circle is moved to the top-left corner of the screen. It's almost as if the Context2D transform applied at the start of the drawing is forgotten partway through. ![image](https://user-images.githubusercontent.com/544744/33791176-e8cd31fc-dc3c-11e7-911a-25e72d8cf13d.png)
process
transform on canvas is spuriously lost the program at acts oddly it should draw an approximate circle out of random triangles instead at least on chrome part of the circle is moved to the top left corner of the screen it s almost as if the transform applied at the start of the drawing is forgotten partway through
1
6,340
9,380,269,240
IssuesEvent
2019-04-04 16:39:38
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
Eldoc output dumped to process buffer
process symbols:completion
ess-20181206.2103 from elpa It seems print some debug information in the inferior buffer as below. The line deparsing the input is inserted into the buffer. ``` > (list "" '(("..." . "")) '("recursive" "use.names" "..." "x")) > (list "" '(("..." . "")) '("recursive" "use.names" "..." "x")) > (list "" '(("..." . "")) '("recursive" "use.names" "..." "x")) > c("RFSTDTC", "BRTHDTC") ```
1.0
Eldoc output dumped to process buffer - ess-20181206.2103 from elpa It seems print some debug information in the inferior buffer as below. The line deparsing the input is inserted into the buffer. ``` > (list "" '(("..." . "")) '("recursive" "use.names" "..." "x")) > (list "" '(("..." . "")) '("recursive" "use.names" "..." "x")) > (list "" '(("..." . "")) '("recursive" "use.names" "..." "x")) > c("RFSTDTC", "BRTHDTC") ```
process
eldoc output dumped to process buffer ess from elpa it seems print some debug information in the inferior buffer as below the line deparsing the input is inserted into the buffer list recursive use names x list recursive use names x list recursive use names x c rfstdtc brthdtc
1
372,619
26,010,675,298
IssuesEvent
2022-12-21 01:10:05
mapbox/dr-ui
https://api.github.com/repos/mapbox/dr-ui
closed
Make sure propTypes comments use `/**` notation
good first issue documentation
To document props on our catalog site, every prop should lead with a comment that starts with double stars: ```jsx ToggleableCodeBlock.propTypes = { /** A unique `id` is required for the language toggle. */ id: PropTypes.string.isRequired } ``` We should make sure every component's propTypes: - Have a comment to describe it on the line above. - Uses the double star to start the comment.
1.0
Make sure propTypes comments use `/**` notation - To document props on our catalog site, every prop should lead with a comment that starts with double stars: ```jsx ToggleableCodeBlock.propTypes = { /** A unique `id` is required for the language toggle. */ id: PropTypes.string.isRequired } ``` We should make sure every component's propTypes: - Have a comment to describe it on the line above. - Uses the double star to start the comment.
non_process
make sure proptypes comments use notation to document props on our catalog site every prop should lead with a comment that starts with double stars jsx toggleablecodeblock proptypes a unique id is required for the language toggle id proptypes string isrequired we should make sure every component s proptypes have a comment to describe it on the line above uses the double star to start the comment
0
6,158
9,037,760,467
IssuesEvent
2019-02-09 14:00:00
kmycode/sangokukmy
https://api.github.com/repos/kmycode/sangokukmy
closed
兵種
enhancement priority-high process-processing
三国志NET KMY Versionでは、兵種を以下のようにしようと考えています。 ## 兵種一覧 攻撃力、防御力は、いずれも武力に加算する値。 ゲームが進行すると、年の1000分の1のボーナスがつく(例:開始300年目の場合、攻撃力と防御力を1.3倍した値と武力の合計が適用される) ※この表は頻繁に見直したり更新したりするのでちうい>< ※雑兵・禁兵は、コマンド入力時ではなくコマンド実行時の都市を見ます。いずれも、コマンド入力時は雑兵を選択します | 種類 | 攻撃力 | 防御力 | 金 | 技術 | 備考 | | -- | -- | -- | -- | -- | -- | | 雑兵 | 0 | 0 | 10 | 0 | 実行時都市が首都の場合は禁兵が徴兵される | | 禁兵 | 20 | 20 | 10 | 0 | 雑兵徴兵コマを首都で実行する必要がある | | 軽歩兵 | 10 | 0 | 20 | 101 | | | 弓兵 | 0 | 15 | 30 | 201 | | | 軽騎兵 | 35 | 10 | 50 | 301 | | | 強弩兵 | 10 | 35 | 70 | 401 | | | 神鬼兵 | 知力 | 0 | 100 | 501 | | | 重歩兵 | 50 | 30 | 150 | 601 | | | 重騎兵 | 60 | 40 | 200 | 701 | | | 智攻兵 | 知力x0.7 | 知力x0.4 | 250 | 801 | | | 連弩兵 | 80 | 10 | 300 | 901 | | 壁守兵 | 0 | 知力 | 350 | 999 | | | 衝車 | 20 | 20 | 200 | 999 | 相手が城壁の時に攻撃力200。#48 へ移動 | | 井闌 | 20 | 20 | 200 | 999 | 相手が守兵/壁守兵の時に攻撃力200、反撃僅少。 #48 へ移動 | ## 徴兵コマンドを入力可能な都市( #47 へ移動) 首都のルールが最優先で適用される。農業都市が首都だった場合、農業都市で✕の兵種も徴兵コマンドを入力できる | 種類 | 農業都市 | 商業都市 | 城塞都市 | 大都市 | 首都 | | --- | :---: | :---: | :---: | :---: | :---: | | 雑兵 | ◯ | ◯ | ◯ | ◯ | ✕ | | 禁兵 | ✕ | ✕ | ✕ | ✕ | ◯ | | 軽歩兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 弓兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 軽騎兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 強弩兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 神鬼兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 重歩兵 | ◯ | ✕ | ◯ | ◯ | ◯ | | 重騎兵 | ◯ | ✕ | ◯ | ◯ | ◯ | | 智攻兵 | ✕ | ◯ | ◯ | ◯ | ◯ | | 連弩兵 | ◯ | ✕ | ◯ | ✕ | ◯ | | 壁守兵 | ✕ | ✕ | ◯ | ◯ | ◯ | | 衝車 | ◯ | ✕ | ✕ | ◯ | ◯ | | 井闌 | ✕ | ◯ | ✕ | ◯ | ◯ | ## 以前との違い 以前の兵種は、ここにまとめられています。 https://www22.atwiki.jp/kmy-sangokusi/pages/11.html * 加年ボーナスがつきました(詳細は上記) * 一部兵種が削減され、主要能力は武力と知力のみになりました * 衝車、井闌が追加されました * 一部兵種のバランスが調整されました * 都市特化による徴兵の可否要素が追加されました * 以前は徴兵時の技術は実行時判定でしたが、コマンド入力時の技術力で判定するよう戻しました
2.0
兵種 - 三国志NET KMY Versionでは、兵種を以下のようにしようと考えています。 ## 兵種一覧 攻撃力、防御力は、いずれも武力に加算する値。 ゲームが進行すると、年の1000分の1のボーナスがつく(例:開始300年目の場合、攻撃力と防御力を1.3倍した値と武力の合計が適用される) ※この表は頻繁に見直したり更新したりするのでちうい>< ※雑兵・禁兵は、コマンド入力時ではなくコマンド実行時の都市を見ます。いずれも、コマンド入力時は雑兵を選択します | 種類 | 攻撃力 | 防御力 | 金 | 技術 | 備考 | | -- | -- | -- | -- | -- | -- | | 雑兵 | 0 | 0 | 10 | 0 | 実行時都市が首都の場合は禁兵が徴兵される | | 禁兵 | 20 | 20 | 10 | 0 | 雑兵徴兵コマを首都で実行する必要がある | | 軽歩兵 | 10 | 0 | 20 | 101 | | | 弓兵 | 0 | 15 | 30 | 201 | | | 軽騎兵 | 35 | 10 | 50 | 301 | | | 強弩兵 | 10 | 35 | 70 | 401 | | | 神鬼兵 | 知力 | 0 | 100 | 501 | | | 重歩兵 | 50 | 30 | 150 | 601 | | | 重騎兵 | 60 | 40 | 200 | 701 | | | 智攻兵 | 知力x0.7 | 知力x0.4 | 250 | 801 | | | 連弩兵 | 80 | 10 | 300 | 901 | | 壁守兵 | 0 | 知力 | 350 | 999 | | | 衝車 | 20 | 20 | 200 | 999 | 相手が城壁の時に攻撃力200。#48 へ移動 | | 井闌 | 20 | 20 | 200 | 999 | 相手が守兵/壁守兵の時に攻撃力200、反撃僅少。 #48 へ移動 | ## 徴兵コマンドを入力可能な都市( #47 へ移動) 首都のルールが最優先で適用される。農業都市が首都だった場合、農業都市で✕の兵種も徴兵コマンドを入力できる | 種類 | 農業都市 | 商業都市 | 城塞都市 | 大都市 | 首都 | | --- | :---: | :---: | :---: | :---: | :---: | | 雑兵 | ◯ | ◯ | ◯ | ◯ | ✕ | | 禁兵 | ✕ | ✕ | ✕ | ✕ | ◯ | | 軽歩兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 弓兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 軽騎兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 強弩兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 神鬼兵 | ◯ | ◯ | ◯ | ◯ | ◯ | | 重歩兵 | ◯ | ✕ | ◯ | ◯ | ◯ | | 重騎兵 | ◯ | ✕ | ◯ | ◯ | ◯ | | 智攻兵 | ✕ | ◯ | ◯ | ◯ | ◯ | | 連弩兵 | ◯ | ✕ | ◯ | ✕ | ◯ | | 壁守兵 | ✕ | ✕ | ◯ | ◯ | ◯ | | 衝車 | ◯ | ✕ | ✕ | ◯ | ◯ | | 井闌 | ✕ | ◯ | ✕ | ◯ | ◯ | ## 以前との違い 以前の兵種は、ここにまとめられています。 https://www22.atwiki.jp/kmy-sangokusi/pages/11.html * 加年ボーナスがつきました(詳細は上記) * 一部兵種が削減され、主要能力は武力と知力のみになりました * 衝車、井闌が追加されました * 一部兵種のバランスが調整されました * 都市特化による徴兵の可否要素が追加されました * 以前は徴兵時の技術は実行時判定でしたが、コマンド入力時の技術力で判定するよう戻しました
process
兵種 三国志net kmy versionでは、兵種を以下のようにしようと考えています。 兵種一覧 攻撃力、防御力は、いずれも武力に加算する値。 ゲームが進行すると、 (例: 、 ) ※この表は頻繁に見直したり更新したりするのでちうい>< ※雑兵・禁兵は、コマンド入力時ではなくコマンド実行時の都市を見ます。いずれも、コマンド入力時は雑兵を選択します 種類 攻撃力 防御力 金 技術 備考 雑兵 実行時都市が首都の場合は禁兵が徴兵される 禁兵 雑兵徴兵コマを首都で実行する必要がある 軽歩兵 弓兵 軽騎兵 強弩兵 神鬼兵 知力 重歩兵 重騎兵 智攻兵 連弩兵 壁守兵 知力 衝車 。 へ移動 井闌 相手が守兵/ 、反撃僅少。 へ移動 徴兵コマンドを入力可能な都市( へ移動) 首都のルールが最優先で適用される。農業都市が首都だった場合、農業都市で✕の兵種も徴兵コマンドを入力できる 種類 農業都市 商業都市 城塞都市 大都市 首都 雑兵 ◯ ◯ ◯ ◯ ✕ 禁兵 ✕ ✕ ✕ ✕ ◯ 軽歩兵 ◯ ◯ ◯ ◯ ◯ 弓兵 ◯ ◯ ◯ ◯ ◯ 軽騎兵 ◯ ◯ ◯ ◯ ◯ 強弩兵 ◯ ◯ ◯ ◯ ◯ 神鬼兵 ◯ ◯ ◯ ◯ ◯ 重歩兵 ◯ ✕ ◯ ◯ ◯ 重騎兵 ◯ ✕ ◯ ◯ ◯ 智攻兵 ✕ ◯ ◯ ◯ ◯ 連弩兵 ◯ ✕ ◯ ✕ ◯ 壁守兵 ✕ ✕ ◯ ◯ ◯ 衝車 ◯ ✕ ✕ ◯ ◯ 井闌 ✕ ◯ ✕ ◯ ◯ 以前との違い 以前の兵種は、ここにまとめられています。 加年ボーナスがつきました(詳細は上記) 一部兵種が削減され、主要能力は武力と知力のみになりました 衝車、井闌が追加されました 一部兵種のバランスが調整されました 都市特化による徴兵の可否要素が追加されました 以前は徴兵時の技術は実行時判定でしたが、コマンド入力時の技術力で判定するよう戻しました
1
798,584
28,290,562,338
IssuesEvent
2023-04-09 06:15:57
7-lin/Final_Project_BE
https://api.github.com/repos/7-lin/Final_Project_BE
closed
feat: 취소된 예약을 다시 되돌리는 기능을 구현한다.
For : API For : Backend Priority : Medium Status : In Progress Type : Feature
## Description(설명) ## Tasks(New feature) - [x] 취소된 예약을 가져와서 엔티티의 값을 WAITING로 바꾸는 로직을 구현한다. - [x] 이미 취소된 예약이라면 취소할수 없다는 response를 내보낸다. - [x] 예약 상세 페이지에 인원수 정보도 나올 수 있도록 수정한다. ## References [link text](link addr)
1.0
feat: 취소된 예약을 다시 되돌리는 기능을 구현한다. - ## Description(설명) ## Tasks(New feature) - [x] 취소된 예약을 가져와서 엔티티의 값을 WAITING로 바꾸는 로직을 구현한다. - [x] 이미 취소된 예약이라면 취소할수 없다는 response를 내보낸다. - [x] 예약 상세 페이지에 인원수 정보도 나올 수 있도록 수정한다. ## References [link text](link addr)
non_process
feat 취소된 예약을 다시 되돌리는 기능을 구현한다 description 설명 tasks new feature 취소된 예약을 가져와서 엔티티의 값을 waiting로 바꾸는 로직을 구현한다 이미 취소된 예약이라면 취소할수 없다는 response를 내보낸다 예약 상세 페이지에 인원수 정보도 나올 수 있도록 수정한다 references link addr
0
3,997
6,926,119,895
IssuesEvent
2017-11-30 18:02:43
ouh-churchill/COPE
https://api.github.com/repos/ouh-churchill/COPE
opened
DMC Modifications Tasklist (Overview task)
App: Systemwide Category: enhancement Priority: High Process: New
Following the meeting on 30th Nov 2017, the following (reduced) overview was minuted (see email at 14:48 for full text): SK: I met with @AllyBradley , @marshalc and LD this morning in order to better understand the WP4 data, issues and come up with a way forward to create robust DMC reports and trial outcome data moving forwards. We spent over 3 hours looking over the data and reports and have come up with a (hopefully foolproof) plan. The first thing that is evident is that there are different data sources being used. Unless data has been directly entered into the database or received on a paper CRF or AE form, it cannot be considered in the trial. We currently have database records for XXX randomised donors (YYY kidneys). This will form the starting point for all analyses/reporting. CM will then be able to provide data for all kidneys that have been excluded, and the reasons for these exclusions, to give us our final number of included kidneys to be analysed. He will report the overall number of kidneys, and the number of complete pairs available. Potential reasons for exclusion after randomisation are: - Not meeting inclusion criteria (no consent, multiorgan/dual transplant) - Kidney not allocated to a site - Kidney discarded after allocation - Allocated to non-participating site - Reallocated to a non-participating site - No evidence of transplant available in the database - No timings available (knife to skin, anastomosis start, reperfusion, operation concluded) AND no f/u forms received It is possible that the final group (no evidence of a transplant in the database) includes some kidneys that were transplanted but we have not received data. Once we have this list, we will provide it for the PI to investigate whether these kidneys were indeed transplanted, and if so, why we have no data available. If necessary, CM can provide trial IDs for each of these groups so that the master database can be compared to any other data sources that you may be keeping. The consort flowchart for the final DMC report will then follow the template attached. LD will produce this once CM has provided the data. We have agreed the following action points: 1. CM will merge AB's records of received CRFs to the database, so that we can easily identify which paper forms we have received, where data were directly entered into the database, and where follow-up is missing. Moving forward, when paper CRFs are transcribed or data are entered into the database, a field will be set to define the data source so that we can report missing data by visit and by patient. 2. CM will write rules to define excluded kidneys and the reasons for exclusion, so that LD can easily see which kidneys are to be included for analysis. 3. CM will flatten the data into a single spreadsheet for LD to analyse. This spreadsheet will have YYY rows - one for each kidney randomised - and include (where applicable/available) - Trial ID - Included/excluded and reason for exclusion - All donor demographics - All procurement detail - All recipient demographics - All transplant procedure data - All follow-up data for each timepoint - Total number of AEs/SAEs for that kidney - Recipient death and date of death (for included kidneys) - defined by death reported on an SAE form - Permanent graft failure (for included kidneys) - defined as report of permanent failure on an SAE form, or graft failure/graft removal on a follow-up form - Delayed graft function (for included kidneys) - defined as dialysis on days 1-7 post-transplant or selection of a reason for dialysis (but not if graft reported as removed on initial follow-up form) 4. LD will use this dataset to produce the DMC report. This should follow the format used by WP3, and does not need to repeat protocol information. AB/SK will review the report and identify any discrepant data. These will be chased and corrected in the master database prior to submission of the report to the DMC.
1.0
DMC Modifications Tasklist (Overview task) - Following the meeting on 30th Nov 2017, the following (reduced) overview was minuted (see email at 14:48 for full text): SK: I met with @AllyBradley , @marshalc and LD this morning in order to better understand the WP4 data, issues and come up with a way forward to create robust DMC reports and trial outcome data moving forwards. We spent over 3 hours looking over the data and reports and have come up with a (hopefully foolproof) plan. The first thing that is evident is that there are different data sources being used. Unless data has been directly entered into the database or received on a paper CRF or AE form, it cannot be considered in the trial. We currently have database records for XXX randomised donors (YYY kidneys). This will form the starting point for all analyses/reporting. CM will then be able to provide data for all kidneys that have been excluded, and the reasons for these exclusions, to give us our final number of included kidneys to be analysed. He will report the overall number of kidneys, and the number of complete pairs available. Potential reasons for exclusion after randomisation are: - Not meeting inclusion criteria (no consent, multiorgan/dual transplant) - Kidney not allocated to a site - Kidney discarded after allocation - Allocated to non-participating site - Reallocated to a non-participating site - No evidence of transplant available in the database - No timings available (knife to skin, anastomosis start, reperfusion, operation concluded) AND no f/u forms received It is possible that the final group (no evidence of a transplant in the database) includes some kidneys that were transplanted but we have not received data. Once we have this list, we will provide it for the PI to investigate whether these kidneys were indeed transplanted, and if so, why we have no data available. If necessary, CM can provide trial IDs for each of these groups so that the master database can be compared to any other data sources that you may be keeping. The consort flowchart for the final DMC report will then follow the template attached. LD will produce this once CM has provided the data. We have agreed the following action points: 1. CM will merge AB's records of received CRFs to the database, so that we can easily identify which paper forms we have received, where data were directly entered into the database, and where follow-up is missing. Moving forward, when paper CRFs are transcribed or data are entered into the database, a field will be set to define the data source so that we can report missing data by visit and by patient. 2. CM will write rules to define excluded kidneys and the reasons for exclusion, so that LD can easily see which kidneys are to be included for analysis. 3. CM will flatten the data into a single spreadsheet for LD to analyse. This spreadsheet will have YYY rows - one for each kidney randomised - and include (where applicable/available) - Trial ID - Included/excluded and reason for exclusion - All donor demographics - All procurement detail - All recipient demographics - All transplant procedure data - All follow-up data for each timepoint - Total number of AEs/SAEs for that kidney - Recipient death and date of death (for included kidneys) - defined by death reported on an SAE form - Permanent graft failure (for included kidneys) - defined as report of permanent failure on an SAE form, or graft failure/graft removal on a follow-up form - Delayed graft function (for included kidneys) - defined as dialysis on days 1-7 post-transplant or selection of a reason for dialysis (but not if graft reported as removed on initial follow-up form) 4. LD will use this dataset to produce the DMC report. This should follow the format used by WP3, and does not need to repeat protocol information. AB/SK will review the report and identify any discrepant data. These will be chased and corrected in the master database prior to submission of the report to the DMC.
process
dmc modifications tasklist overview task following the meeting on nov the following reduced overview was minuted see email at for full text sk i met with allybradley marshalc and ld this morning in order to better understand the data issues and come up with a way forward to create robust dmc reports and trial outcome data moving forwards we spent over hours looking over the data and reports and have come up with a hopefully foolproof plan the first thing that is evident is that there are different data sources being used unless data has been directly entered into the database or received on a paper crf or ae form it cannot be considered in the trial we currently have database records for xxx randomised donors yyy kidneys this will form the starting point for all analyses reporting cm will then be able to provide data for all kidneys that have been excluded and the reasons for these exclusions to give us our final number of included kidneys to be analysed he will report the overall number of kidneys and the number of complete pairs available potential reasons for exclusion after randomisation are not meeting inclusion criteria no consent multiorgan dual transplant kidney not allocated to a site kidney discarded after allocation allocated to non participating site reallocated to a non participating site no evidence of transplant available in the database no timings available knife to skin anastomosis start reperfusion operation concluded and no f u forms received it is possible that the final group no evidence of a transplant in the database includes some kidneys that were transplanted but we have not received data once we have this list we will provide it for the pi to investigate whether these kidneys were indeed transplanted and if so why we have no data available if necessary cm can provide trial ids for each of these groups so that the master database can be compared to any other data sources that you may be keeping the consort flowchart for the final dmc report will then follow the template attached ld will produce this once cm has provided the data we have agreed the following action points cm will merge ab s records of received crfs to the database so that we can easily identify which paper forms we have received where data were directly entered into the database and where follow up is missing moving forward when paper crfs are transcribed or data are entered into the database a field will be set to define the data source so that we can report missing data by visit and by patient cm will write rules to define excluded kidneys and the reasons for exclusion so that ld can easily see which kidneys are to be included for analysis cm will flatten the data into a single spreadsheet for ld to analyse this spreadsheet will have yyy rows one for each kidney randomised and include where applicable available trial id included excluded and reason for exclusion all donor demographics all procurement detail all recipient demographics all transplant procedure data all follow up data for each timepoint total number of aes saes for that kidney recipient death and date of death for included kidneys defined by death reported on an sae form permanent graft failure for included kidneys defined as report of permanent failure on an sae form or graft failure graft removal on a follow up form delayed graft function for included kidneys defined as dialysis on days post transplant or selection of a reason for dialysis but not if graft reported as removed on initial follow up form ld will use this dataset to produce the dmc report this should follow the format used by and does not need to repeat protocol information ab sk will review the report and identify any discrepant data these will be chased and corrected in the master database prior to submission of the report to the dmc
1
8,475
4,254,961,543
IssuesEvent
2016-07-09 04:20:09
jeff1evesque/machine-learning
https://api.github.com/repos/jeff1evesque/machine-learning
closed
Dynamically define 'dev_env_path' in compiler manifests
bug build
We need to change all our existing compiler manifests, from both of our existing puppet `environment`: ```bash ... $dev_env_path = "${root_dir}/puppet/environment/development" ... ``` to something of the form: ```bash ... $environment = $hiera_general['environment'] $dev_env_path = "${root_dir}/puppet/environment/${environment}" ... ```
1.0
Dynamically define 'dev_env_path' in compiler manifests - We need to change all our existing compiler manifests, from both of our existing puppet `environment`: ```bash ... $dev_env_path = "${root_dir}/puppet/environment/development" ... ``` to something of the form: ```bash ... $environment = $hiera_general['environment'] $dev_env_path = "${root_dir}/puppet/environment/${environment}" ... ```
non_process
dynamically define dev env path in compiler manifests we need to change all our existing compiler manifests from both of our existing puppet environment bash dev env path root dir puppet environment development to something of the form bash environment hiera general dev env path root dir puppet environment environment
0
82,153
15,646,513,990
IssuesEvent
2021-03-23 01:06:05
jgeraigery/linux
https://api.github.com/repos/jgeraigery/linux
opened
CVE-2020-27152 (Medium) detected in linuxv5.2
security vulnerability
## CVE-2020-27152 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ioapic_lazy_update_eoi in arch/x86/kvm/ioapic.c in the Linux kernel before 5.9.2. It has an infinite loop related to improper interaction between a resampler and edge triggering, aka CID-77377064c3a9. <p>Publish Date: 2020-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27152>CVE-2020-27152</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/gregkh/linux/releases/tag/v5.9.2">https://github.com/gregkh/linux/releases/tag/v5.9.2</a></p> <p>Release Date: 2020-10-15</p> <p>Fix Resolution: v5.9.2</p> </p> </details> <p></p>
True
CVE-2020-27152 (Medium) detected in linuxv5.2 - ## CVE-2020-27152 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ioapic_lazy_update_eoi in arch/x86/kvm/ioapic.c in the Linux kernel before 5.9.2. It has an infinite loop related to improper interaction between a resampler and edge triggering, aka CID-77377064c3a9. <p>Publish Date: 2020-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27152>CVE-2020-27152</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/gregkh/linux/releases/tag/v5.9.2">https://github.com/gregkh/linux/releases/tag/v5.9.2</a></p> <p>Release Date: 2020-10-15</p> <p>Fix Resolution: v5.9.2</p> </p> </details> <p></p>
non_process
cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href vulnerable source files vulnerability details an issue was discovered in ioapic lazy update eoi in arch kvm ioapic c in the linux kernel before it has an infinite loop related to improper interaction between a resampler and edge triggering aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
133,615
12,546,300,098
IssuesEvent
2020-06-05 20:27:17
ZupIT/charlescd
https://api.github.com/repos/ZupIT/charlescd
closed
Broken link in docs - section getting started
bug documentation
**Describe the bug** Broken link in docs - section getting started **To Reproduce** Steps to reproduce the behavior: 1. Go to https://docs.charlescd.io/v/v0.2.1/primeiros-passos/instalando-charles 2. Click in "customizacão" ( https://docs.charlescd.io/primeiros-passos/instalando-charles#customizacao-total ) 3. See error **Expected behavior** Show the page **Screenshots** **Your Environment** - CharlesCD version used: - Description of environment where CharlesCD is running: - Browser Name and version (if applicable): **Additional context** Add any other context about the problem here.
1.0
Broken link in docs - section getting started - **Describe the bug** Broken link in docs - section getting started **To Reproduce** Steps to reproduce the behavior: 1. Go to https://docs.charlescd.io/v/v0.2.1/primeiros-passos/instalando-charles 2. Click in "customizacão" ( https://docs.charlescd.io/primeiros-passos/instalando-charles#customizacao-total ) 3. See error **Expected behavior** Show the page **Screenshots** **Your Environment** - CharlesCD version used: - Description of environment where CharlesCD is running: - Browser Name and version (if applicable): **Additional context** Add any other context about the problem here.
non_process
broken link in docs section getting started describe the bug broken link in docs section getting started to reproduce steps to reproduce the behavior go to click in customizacão see error expected behavior show the page screenshots your environment charlescd version used description of environment where charlescd is running browser name and version if applicable additional context add any other context about the problem here
0
7,102
10,256,710,403
IssuesEvent
2019-08-21 18:19:50
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Track certain metrics over time
process: tests stage: needs review type: chore
We should do some long-term logging of various statistics that are collected about Cypress during CI builds. Things that would be nice to track on each commit: - results of `proxy_performance` server spec - results of `6_visit_spec` "resolves visits quickly": https://github.com/cypress-io/cypress/blob/c4bf698013998a38ff955f5487275ed47e41c5fd/packages/server/test/e2e/6_visit_spec.coffee#L155-L155
1.0
Track certain metrics over time - We should do some long-term logging of various statistics that are collected about Cypress during CI builds. Things that would be nice to track on each commit: - results of `proxy_performance` server spec - results of `6_visit_spec` "resolves visits quickly": https://github.com/cypress-io/cypress/blob/c4bf698013998a38ff955f5487275ed47e41c5fd/packages/server/test/e2e/6_visit_spec.coffee#L155-L155
process
track certain metrics over time we should do some long term logging of various statistics that are collected about cypress during ci builds things that would be nice to track on each commit results of proxy performance server spec results of visit spec resolves visits quickly
1
2,890
5,871,319,842
IssuesEvent
2017-05-15 08:23:00
MobileOrg/mobileorg
https://api.github.com/repos/MobileOrg/mobileorg
closed
Representative Testfiles
development process
To be able to test during development and beta-test. - Should be pretty complex. - Should be shared with developers, testers and users - Should include org-mode configuration (as some defects could occur on emacs, e.g. #97)
1.0
Representative Testfiles - To be able to test during development and beta-test. - Should be pretty complex. - Should be shared with developers, testers and users - Should include org-mode configuration (as some defects could occur on emacs, e.g. #97)
process
representative testfiles to be able to test during development and beta test should be pretty complex should be shared with developers testers and users should include org mode configuration as some defects could occur on emacs e g
1
579,235
17,186,164,356
IssuesEvent
2021-07-16 02:29:44
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
Define Previewable page scope
Q321-priority frontend frontend-vamc vsa vsa-facilities
## Issue Description A gap exists in the front end experience between the preview server and the production environment, making troubleshooting difficult and inefficient for engineers and hard for field staff to view/trust their changes before publishing. This impacts Vet Centers and VAMC --- ## Tasks - [ ] Document the scope of work for making Vet Center and VAMC pages "previewable", including the following considerations: - Audit existing Facilitiies pages to make sure they are all previewable. - Pages with greatest difference between preview and prod, such as entire pages and or areas of missing content (i..e images, alerts) - Document pain point for engineers - troubleshooting bugs, etc (most will be the same) - Value to content editors and business stakeholders - ability to preview content before publishing - Pages at highest risk for bugs/high volume changes by field - Known differences in Preview mode that we want to keep ## Acceptance Criteria - [ ] Scope of initiative is documented and time boxed to pages with greatest value to engineers and users
1.0
Define Previewable page scope - ## Issue Description A gap exists in the front end experience between the preview server and the production environment, making troubleshooting difficult and inefficient for engineers and hard for field staff to view/trust their changes before publishing. This impacts Vet Centers and VAMC --- ## Tasks - [ ] Document the scope of work for making Vet Center and VAMC pages "previewable", including the following considerations: - Audit existing Facilitiies pages to make sure they are all previewable. - Pages with greatest difference between preview and prod, such as entire pages and or areas of missing content (i..e images, alerts) - Document pain point for engineers - troubleshooting bugs, etc (most will be the same) - Value to content editors and business stakeholders - ability to preview content before publishing - Pages at highest risk for bugs/high volume changes by field - Known differences in Preview mode that we want to keep ## Acceptance Criteria - [ ] Scope of initiative is documented and time boxed to pages with greatest value to engineers and users
non_process
define previewable page scope issue description a gap exists in the front end experience between the preview server and the production environment making troubleshooting difficult and inefficient for engineers and hard for field staff to view trust their changes before publishing this impacts vet centers and vamc tasks document the scope of work for making vet center and vamc pages previewable including the following considerations audit existing facilitiies pages to make sure they are all previewable pages with greatest difference between preview and prod such as entire pages and or areas of missing content i e images alerts document pain point for engineers troubleshooting bugs etc most will be the same value to content editors and business stakeholders ability to preview content before publishing pages at highest risk for bugs high volume changes by field known differences in preview mode that we want to keep acceptance criteria scope of initiative is documented and time boxed to pages with greatest value to engineers and users
0
17,389
23,207,325,743
IssuesEvent
2022-08-02 07:01:08
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Bazel does not provide the right path to runfiles in c++
type: support / not a bug (process) team-Rules-CPP untriaged
### Description of the bug: Im trying to access runfiles within c++. Im using Bazel 5.2.0. I tried to access like this: ``` std::string error; std::unique_ptr<Runfiles> runfiles(Runfiles::Create(argv[0], &error)); if (!runfiles) { std::cerr << error << std::endl; return 1; } std::string path = runfiles->Rlocation("Test/Example.tx"); std::cout << "Example.tx: " << path << std::endl; std::ifstream in(path); if (!in.is_open()) { std::cout << "Example.tx not found" << std::endl; return -1; } ``` (Example.tx is right, just to lazy to change)<br/> The program is finding a path but the path starts from the bazelisk directory and doesn't point to the binary dir. ``` Example.tx: C:\users\nikla\_bazel_nikla\d47dtf2d\execroot\__main__\bazel-out\x64_windows-fastbuild\bin\Test\Test.exe.runfiles/Test/Example.tx Example.tx not found ``` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. using "@bazel_tools//tools/cpp/runfiles" as dependency on Windows to access runfiles (using --enable_runfiles). Running with "bazel run --enable_runfiles //Test:Test" ### Which operating system are you running Bazel on? Windows 10 ### What is the output of `bazel info release`? release 6.0.0-pre.20220706.4 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? ![Path](https://user-images.githubusercontent.com/68678245/182185322-69485b8f-2adb-468d-a114-4a39694f4052.PNG) Maybe the run command should run this .exe but just an spontaneous thought.
1.0
Bazel does not provide the right path to runfiles in c++ - ### Description of the bug: Im trying to access runfiles within c++. Im using Bazel 5.2.0. I tried to access like this: ``` std::string error; std::unique_ptr<Runfiles> runfiles(Runfiles::Create(argv[0], &error)); if (!runfiles) { std::cerr << error << std::endl; return 1; } std::string path = runfiles->Rlocation("Test/Example.tx"); std::cout << "Example.tx: " << path << std::endl; std::ifstream in(path); if (!in.is_open()) { std::cout << "Example.tx not found" << std::endl; return -1; } ``` (Example.tx is right, just to lazy to change)<br/> The program is finding a path but the path starts from the bazelisk directory and doesn't point to the binary dir. ``` Example.tx: C:\users\nikla\_bazel_nikla\d47dtf2d\execroot\__main__\bazel-out\x64_windows-fastbuild\bin\Test\Test.exe.runfiles/Test/Example.tx Example.tx not found ``` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. using "@bazel_tools//tools/cpp/runfiles" as dependency on Windows to access runfiles (using --enable_runfiles). Running with "bazel run --enable_runfiles //Test:Test" ### Which operating system are you running Bazel on? Windows 10 ### What is the output of `bazel info release`? release 6.0.0-pre.20220706.4 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? ![Path](https://user-images.githubusercontent.com/68678245/182185322-69485b8f-2adb-468d-a114-4a39694f4052.PNG) Maybe the run command should run this .exe but just an spontaneous thought.
process
bazel does not provide the right path to runfiles in c description of the bug im trying to access runfiles within c im using bazel i tried to access like this std string error std unique ptr runfiles runfiles create argv error if runfiles std cerr error std endl return std string path runfiles rlocation test example tx std cout example tx path std endl std ifstream in path if in is open std cout example tx not found std endl return example tx is right just to lazy to change the program is finding a path but the path starts from the bazelisk directory and doesn t point to the binary dir example tx c users nikla bazel nikla execroot main bazel out windows fastbuild bin test test exe runfiles test example tx example tx not found what s the simplest easiest way to reproduce this bug please provide a minimal example if possible using bazel tools tools cpp runfiles as dependency on windows to access runfiles using enable runfiles running with bazel run enable runfiles test test which operating system are you running bazel on windows what is the output of bazel info release release pre if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no response any other information logs or outputs that you want to share maybe the run command should run this exe but just an spontaneous thought
1
9,706
12,703,396,209
IssuesEvent
2020-06-22 22:13:06
Arch666Angel/mods
https://api.github.com/repos/Arch666Angel/mods
closed
[Enhancement] Make map colors of gardens and puffer nests distinguishable.
Angels Bio Processing Impact: Enhancement
Currently, revealing the map with radars will show locations of gardens, puffer nests, and the "special" trees. Trees have a deep blue color, but gardens and puffer nests are practically indistinuishable. To be honest, I am not sure if they have the same color or slightly different – and I just cannot see the difference. Please make gardens and puffer nests use different colors on the map. This makes exploring and hunting for gardens less tedious. ![garden_puffer-nest](https://user-images.githubusercontent.com/1843557/84349048-7b3f4c00-ab84-11ea-8357-70948965c1f6.jpg)
1.0
[Enhancement] Make map colors of gardens and puffer nests distinguishable. - Currently, revealing the map with radars will show locations of gardens, puffer nests, and the "special" trees. Trees have a deep blue color, but gardens and puffer nests are practically indistinuishable. To be honest, I am not sure if they have the same color or slightly different – and I just cannot see the difference. Please make gardens and puffer nests use different colors on the map. This makes exploring and hunting for gardens less tedious. ![garden_puffer-nest](https://user-images.githubusercontent.com/1843557/84349048-7b3f4c00-ab84-11ea-8357-70948965c1f6.jpg)
process
make map colors of gardens and puffer nests distinguishable currently revealing the map with radars will show locations of gardens puffer nests and the special trees trees have a deep blue color but gardens and puffer nests are practically indistinuishable to be honest i am not sure if they have the same color or slightly different – and i just cannot see the difference please make gardens and puffer nests use different colors on the map this makes exploring and hunting for gardens less tedious
1
2,130
4,260,107,303
IssuesEvent
2016-07-11 13:25:56
IBM-Bluemix/logistics-wizard
https://api.github.com/repos/IBM-Bluemix/logistics-wizard
opened
Inventory and shipment management API
erp-service story
Aim of this story is to handle the inventory updates as shipments are created and delivered. The current Inventory model in the simulator has no business logic. _Proposed changes_ * Remove the default Loopback generated API in order to implement custom business logic * Introduce "reserved quantity" in product inventory at a location to avoid shipping the same product twice _Lifecycle of a Shipment_ **To create a new shipment** * Use POST /Shipments to create a new shipment * Use POST /Shipments/{id}/items to add line items to a shipment * Shipment status becomes NEW **To approve a shipment** * Use POST /Shipments/{id}/approve * Shipment status becomes APPROVED * Shipment can no longer be modified * Line items can no longer be modified * Inventory lines for the products in the shipments are updated and quantity reserved. **To notify of shipping in progress** * Use POST /Shipments/{id}/intransit to update location data * Only if current status is APPROVED or IN_TRANSIT * Shipment status becomes IN_TRANSIT **To complete the delivery** * Use POST /Shipments/{id}/delivered * Only if current status is IN_TRANSIT * Shipment status becomes DELIVERED * Inventory lines for the products in the shipments are updated, actual stock is decreased, reserved quantity is updated.
1.0
Inventory and shipment management API - Aim of this story is to handle the inventory updates as shipments are created and delivered. The current Inventory model in the simulator has no business logic. _Proposed changes_ * Remove the default Loopback generated API in order to implement custom business logic * Introduce "reserved quantity" in product inventory at a location to avoid shipping the same product twice _Lifecycle of a Shipment_ **To create a new shipment** * Use POST /Shipments to create a new shipment * Use POST /Shipments/{id}/items to add line items to a shipment * Shipment status becomes NEW **To approve a shipment** * Use POST /Shipments/{id}/approve * Shipment status becomes APPROVED * Shipment can no longer be modified * Line items can no longer be modified * Inventory lines for the products in the shipments are updated and quantity reserved. **To notify of shipping in progress** * Use POST /Shipments/{id}/intransit to update location data * Only if current status is APPROVED or IN_TRANSIT * Shipment status becomes IN_TRANSIT **To complete the delivery** * Use POST /Shipments/{id}/delivered * Only if current status is IN_TRANSIT * Shipment status becomes DELIVERED * Inventory lines for the products in the shipments are updated, actual stock is decreased, reserved quantity is updated.
non_process
inventory and shipment management api aim of this story is to handle the inventory updates as shipments are created and delivered the current inventory model in the simulator has no business logic proposed changes remove the default loopback generated api in order to implement custom business logic introduce reserved quantity in product inventory at a location to avoid shipping the same product twice lifecycle of a shipment to create a new shipment use post shipments to create a new shipment use post shipments id items to add line items to a shipment shipment status becomes new to approve a shipment use post shipments id approve shipment status becomes approved shipment can no longer be modified line items can no longer be modified inventory lines for the products in the shipments are updated and quantity reserved to notify of shipping in progress use post shipments id intransit to update location data only if current status is approved or in transit shipment status becomes in transit to complete the delivery use post shipments id delivered only if current status is in transit shipment status becomes delivered inventory lines for the products in the shipments are updated actual stock is decreased reserved quantity is updated
0
57,187
8,153,999,886
IssuesEvent
2018-08-23 00:45:28
lithnet/resourcemanagement-powershell
https://api.github.com/repos/lithnet/resourcemanagement-powershell
closed
Doc: Possible wrong Xpath in two user examples
documentation
On the example page for Search-Resource (https://github.com/lithnet/resourcemanagement-powershell/wiki/Search-Resources) there are two examples regarding searching for users. The code remarks describe users but the code writes 'Set'. Shouldn't it be /Person instead of /Set for those two examples? ![image](https://user-images.githubusercontent.com/1877612/44391956-5e5e0280-a531-11e8-89be-c52cc1a2378f.png)
1.0
Doc: Possible wrong Xpath in two user examples - On the example page for Search-Resource (https://github.com/lithnet/resourcemanagement-powershell/wiki/Search-Resources) there are two examples regarding searching for users. The code remarks describe users but the code writes 'Set'. Shouldn't it be /Person instead of /Set for those two examples? ![image](https://user-images.githubusercontent.com/1877612/44391956-5e5e0280-a531-11e8-89be-c52cc1a2378f.png)
non_process
doc possible wrong xpath in two user examples on the example page for search resource there are two examples regarding searching for users the code remarks describe users but the code writes set shouldn t it be person instead of set for those two examples
0
123,217
17,772,189,665
IssuesEvent
2021-08-30 14:50:14
kapseliboi/sqlpad
https://api.github.com/repos/kapseliboi/sqlpad
opened
CVE-2021-39171 (Medium) detected in passport-saml-2.2.0.tgz
security vulnerability
## CVE-2021-39171 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>passport-saml-2.2.0.tgz</b></p></summary> <p>SAML 2.0 authentication strategy for Passport</p> <p>Library home page: <a href="https://registry.npmjs.org/passport-saml/-/passport-saml-2.2.0.tgz">https://registry.npmjs.org/passport-saml/-/passport-saml-2.2.0.tgz</a></p> <p>Path to dependency file: sqlpad/server/package.json</p> <p>Path to vulnerable library: /server/node_modules/passport-saml/package.json</p> <p> Dependency Hierarchy: - :x: **passport-saml-2.2.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/sqlpad/commit/95024fc09fd71a1bc52f23bc0709ce5daa7e9f98">95024fc09fd71a1bc52f23bc0709ce5daa7e9f98</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Passport-SAML is a SAML 2.0 authentication provider for Passport, the Node.js authentication library. Prior to version 3.1.0, a malicious SAML payload can require transforms that consume significant system resources to process, thereby resulting in reduced or denied service. This would be an effective way to perform a denial-of-service attack. This has been resolved in version 3.1.0. The resolution is to limit the number of allowable transforms to 2. <p>Publish Date: 2021-08-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-39171>CVE-2021-39171</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/node-saml/passport-saml/security/advisories/GHSA-5379-r78w-42h2">https://github.com/node-saml/passport-saml/security/advisories/GHSA-5379-r78w-42h2</a></p> <p>Release Date: 2021-08-27</p> <p>Fix Resolution: passport-saml - 3.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-39171 (Medium) detected in passport-saml-2.2.0.tgz - ## CVE-2021-39171 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>passport-saml-2.2.0.tgz</b></p></summary> <p>SAML 2.0 authentication strategy for Passport</p> <p>Library home page: <a href="https://registry.npmjs.org/passport-saml/-/passport-saml-2.2.0.tgz">https://registry.npmjs.org/passport-saml/-/passport-saml-2.2.0.tgz</a></p> <p>Path to dependency file: sqlpad/server/package.json</p> <p>Path to vulnerable library: /server/node_modules/passport-saml/package.json</p> <p> Dependency Hierarchy: - :x: **passport-saml-2.2.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/sqlpad/commit/95024fc09fd71a1bc52f23bc0709ce5daa7e9f98">95024fc09fd71a1bc52f23bc0709ce5daa7e9f98</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Passport-SAML is a SAML 2.0 authentication provider for Passport, the Node.js authentication library. Prior to version 3.1.0, a malicious SAML payload can require transforms that consume significant system resources to process, thereby resulting in reduced or denied service. This would be an effective way to perform a denial-of-service attack. This has been resolved in version 3.1.0. The resolution is to limit the number of allowable transforms to 2. <p>Publish Date: 2021-08-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-39171>CVE-2021-39171</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/node-saml/passport-saml/security/advisories/GHSA-5379-r78w-42h2">https://github.com/node-saml/passport-saml/security/advisories/GHSA-5379-r78w-42h2</a></p> <p>Release Date: 2021-08-27</p> <p>Fix Resolution: passport-saml - 3.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in passport saml tgz cve medium severity vulnerability vulnerable library passport saml tgz saml authentication strategy for passport library home page a href path to dependency file sqlpad server package json path to vulnerable library server node modules passport saml package json dependency hierarchy x passport saml tgz vulnerable library found in head commit a href found in base branch master vulnerability details passport saml is a saml authentication provider for passport the node js authentication library prior to version a malicious saml payload can require transforms that consume significant system resources to process thereby resulting in reduced or denied service this would be an effective way to perform a denial of service attack this has been resolved in version the resolution is to limit the number of allowable transforms to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution passport saml step up your open source security game with whitesource
0
56,110
11,504,450,732
IssuesEvent
2020-02-12 23:20:50
microsoft/vscode-python
https://api.github.com/repos/microsoft/vscode-python
closed
Recommend nvm for managing node in CONTRIBUTING.md.
feature-* needs PR type-code health
The extension requires node 8.9.1 (for now), but linux distros may install a newer version (e.g. 8.10 on Ubuntu). You can use `nvm` to manage the node version. The `.travis.yml` file shows what you'll need to do. This should be explained specifically in CONTRIBUTING.md. Also see #1186.
1.0
Recommend nvm for managing node in CONTRIBUTING.md. - The extension requires node 8.9.1 (for now), but linux distros may install a newer version (e.g. 8.10 on Ubuntu). You can use `nvm` to manage the node version. The `.travis.yml` file shows what you'll need to do. This should be explained specifically in CONTRIBUTING.md. Also see #1186.
non_process
recommend nvm for managing node in contributing md the extension requires node for now but linux distros may install a newer version e g on ubuntu you can use nvm to manage the node version the travis yml file shows what you ll need to do this should be explained specifically in contributing md also see
0
6,487
6,474,760,808
IssuesEvent
2017-08-17 18:49:37
zooniverse/caesar
https://api.github.com/repos/zooniverse/caesar
closed
Store caesar configuration in caesar instead of on panoptes API
enhancement infrastructure
Now that the authentication feature has been complete, we can add this feature.
1.0
Store caesar configuration in caesar instead of on panoptes API - Now that the authentication feature has been complete, we can add this feature.
non_process
store caesar configuration in caesar instead of on panoptes api now that the authentication feature has been complete we can add this feature
0
14,190
17,093,839,157
IssuesEvent
2021-07-08 21:34:37
googleapis/python-ndb
https://api.github.com/repos/googleapis/python-ndb
closed
Exercise tests against minimum versions of dependencies
api: datastore type: process
We need to ensure that the package still works against the minimum pinned versions of dependencies. The current pattern for doing so in other repos is to add `testing/constraints-$python_version.txt` files, and have the "oldest" one (`3.6`) exercise the minimum version pins. We can also use `testing/constraints-2.7.txt` to pin 2.7-compatible versions of libraries.
1.0
Exercise tests against minimum versions of dependencies - We need to ensure that the package still works against the minimum pinned versions of dependencies. The current pattern for doing so in other repos is to add `testing/constraints-$python_version.txt` files, and have the "oldest" one (`3.6`) exercise the minimum version pins. We can also use `testing/constraints-2.7.txt` to pin 2.7-compatible versions of libraries.
process
exercise tests against minimum versions of dependencies we need to ensure that the package still works against the minimum pinned versions of dependencies the current pattern for doing so in other repos is to add testing constraints python version txt files and have the oldest one exercise the minimum version pins we can also use testing constraints txt to pin compatible versions of libraries
1
20,042
26,529,381,488
IssuesEvent
2023-01-19 11:15:17
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Introspection of MySQL views
process/candidate topic: introspection topic: re-introspection tech/engines/introspection engine team/schema topic: view kind/subtask
We should fit this beauty to the SQL describer: ```sql select col.table_schema as database_name, col.table_name as view_name, col.ordinal_position, col.column_name, col.data_type, case when col.character_maximum_length is not null then col.character_maximum_length else col.numeric_precision end as max_length, col.is_nullable from information_schema.columns col join information_schema.views vie on vie.table_schema = col.table_schema and vie.table_name = col.table_name where col.table_schema not in ('sys','information_schema', 'mysql', 'performance_schema') -- and vie.table_schema = 'database_name' -- put your database name here order by col.table_schema, col.table_name, col.ordinal_position; ``` There can be relations between views, or models and views. These are not based on foreign keys. They should be kept in re-introspection, for now. Part of: https://github.com/prisma/prisma/issues/17412
1.0
Introspection of MySQL views - We should fit this beauty to the SQL describer: ```sql select col.table_schema as database_name, col.table_name as view_name, col.ordinal_position, col.column_name, col.data_type, case when col.character_maximum_length is not null then col.character_maximum_length else col.numeric_precision end as max_length, col.is_nullable from information_schema.columns col join information_schema.views vie on vie.table_schema = col.table_schema and vie.table_name = col.table_name where col.table_schema not in ('sys','information_schema', 'mysql', 'performance_schema') -- and vie.table_schema = 'database_name' -- put your database name here order by col.table_schema, col.table_name, col.ordinal_position; ``` There can be relations between views, or models and views. These are not based on foreign keys. They should be kept in re-introspection, for now. Part of: https://github.com/prisma/prisma/issues/17412
process
introspection of mysql views we should fit this beauty to the sql describer sql select col table schema as database name col table name as view name col ordinal position col column name col data type case when col character maximum length is not null then col character maximum length else col numeric precision end as max length col is nullable from information schema columns col join information schema views vie on vie table schema col table schema and vie table name col table name where col table schema not in sys information schema mysql performance schema and vie table schema database name put your database name here order by col table schema col table name col ordinal position there can be relations between views or models and views these are not based on foreign keys they should be kept in re introspection for now part of
1
8,248
11,421,369,560
IssuesEvent
2020-02-03 12:02:29
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
.pcss don't work with parcel
:bug: Bug CSS Preprocessing Stale
I setup a project with .pcss files from main HTML. I have `.postcssrc` config file for config some postcss plugins. Parcel serve give me this error when save .pcss file (JS Console): ``` Uncaught Error: Cannot find module '../node_modules/parcel-bundler/src/builtins/css-loader.js' ``` If I change file.pcss to file.css works fine (but I need .pcss for postcss nested syntax and others). If I save html file, works fine also.
1.0
.pcss don't work with parcel - I setup a project with .pcss files from main HTML. I have `.postcssrc` config file for config some postcss plugins. Parcel serve give me this error when save .pcss file (JS Console): ``` Uncaught Error: Cannot find module '../node_modules/parcel-bundler/src/builtins/css-loader.js' ``` If I change file.pcss to file.css works fine (but I need .pcss for postcss nested syntax and others). If I save html file, works fine also.
process
pcss don t work with parcel i setup a project with pcss files from main html i have postcssrc config file for config some postcss plugins parcel serve give me this error when save pcss file js console uncaught error cannot find module node modules parcel bundler src builtins css loader js if i change file pcss to file css works fine but i need pcss for postcss nested syntax and others if i save html file works fine also
1
224,134
7,467,009,764
IssuesEvent
2018-04-02 13:40:16
RileyBee/capstone_project
https://api.github.com/repos/RileyBee/capstone_project
opened
Complete building tables
Priority: Cricital Status: Accepted Status: In Progress Type: Enhancement
While APIs are unavailable for other features of web application, get data from traditional means and build tables for the missing data.
1.0
Complete building tables - While APIs are unavailable for other features of web application, get data from traditional means and build tables for the missing data.
non_process
complete building tables while apis are unavailable for other features of web application get data from traditional means and build tables for the missing data
0
1,962
4,777,699,697
IssuesEvent
2016-10-27 17:01:36
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
uri and path generated incorrectly in .job.xml
bug P2 preprocess
(I'm seeing a bunch of errors like the one I'm going to describe). I am getting the following error: > [gen-list] [DOTX008E][ERROR] File 'file:/Users/tavery/documentation/DITA/source/en/SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita/' does not exist or cannot be loaded. In the ditamap, this topic is referenced as: ```xml <topicref href="SiteGenesis/SiteGenesisSetup.dita"> ``` It is not referenced by any other topic. It exists at the path described. It is valid DITA. I've made sure the header information for the map and topic is identical to my other files (DOCTYPE etc). In the temp files, I'm seeing the following: **In .job.xml it's doubling the file name:** ```xml <file src="file:/Users/tavery/documentation/DITA/source/en/SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita/" uri="SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita/" path="SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita" target="true" non-conref-target="true"/> ``` **In canditopics.list it's also doubling the name:** ``` SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita ``` **In hreftargets.list it's again doubling the name:** ``` SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita ``` I'd attach the files, but the editor won't let me.
1.0
uri and path generated incorrectly in .job.xml - (I'm seeing a bunch of errors like the one I'm going to describe). I am getting the following error: > [gen-list] [DOTX008E][ERROR] File 'file:/Users/tavery/documentation/DITA/source/en/SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita/' does not exist or cannot be loaded. In the ditamap, this topic is referenced as: ```xml <topicref href="SiteGenesis/SiteGenesisSetup.dita"> ``` It is not referenced by any other topic. It exists at the path described. It is valid DITA. I've made sure the header information for the map and topic is identical to my other files (DOCTYPE etc). In the temp files, I'm seeing the following: **In .job.xml it's doubling the file name:** ```xml <file src="file:/Users/tavery/documentation/DITA/source/en/SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita/" uri="SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita/" path="SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita" target="true" non-conref-target="true"/> ``` **In canditopics.list it's also doubling the name:** ``` SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita ``` **In hreftargets.list it's again doubling the name:** ``` SiteGenesis/SiteGenesisSetup.dita/SiteGenesisSetup.dita ``` I'd attach the files, but the editor won't let me.
process
uri and path generated incorrectly in job xml i m seeing a bunch of errors like the one i m going to describe i am getting the following error file file users tavery documentation dita source en sitegenesis sitegenesissetup dita sitegenesissetup dita does not exist or cannot be loaded in the ditamap this topic is referenced as xml it is not referenced by any other topic it exists at the path described it is valid dita i ve made sure the header information for the map and topic is identical to my other files doctype etc in the temp files i m seeing the following in job xml it s doubling the file name xml file src file users tavery documentation dita source en sitegenesis sitegenesissetup dita sitegenesissetup dita uri sitegenesis sitegenesissetup dita sitegenesissetup dita path sitegenesis sitegenesissetup dita sitegenesissetup dita target true non conref target true in canditopics list it s also doubling the name sitegenesis sitegenesissetup dita sitegenesissetup dita in hreftargets list it s again doubling the name sitegenesis sitegenesissetup dita sitegenesissetup dita i d attach the files but the editor won t let me
1
166,893
6,314,176,788
IssuesEvent
2017-07-24 10:06:42
OpenSourceBrain/geppetto-osb
https://api.github.com/repos/OpenSourceBrain/geppetto-osb
closed
Ability to use radius from annotation in NeuroML file for abstract cells
New feature Priority
Many abstract cells in NeuroML2 have no 3D information, e.g. I&F, Izhikevich, specifically nothing saying what radius the should be when rendered in 3D. The default for this in OSB is to use 1um, as this has been useful for the C elegans model, e.g. [here](http://opensourcebrain.org/projects/celegans?explorer=https%253A%252F%252Fraw.githubusercontent.com%252Fopenworm%252FCElegansNeuroML%252Fmaster%252FCElegans%252FpythonScripts%252Fc302%252Fexamples%252Fc302_A_Muscles.nml#) since this appropriate for the radius for somas in the worm. Ideally the default should be 5um, but this could be overridden [in the `<population>` element](https://github.com/openworm/CElegansNeuroML/blob/master/CElegans/pythonScripts/c302/examples/c302_B_Muscles.nml#L59) in the neuroml file inside an `<annotation>` (similar to https://github.com/OpenSourceBrain/redmine/issues/244) allowing different networks to customise how they are displayed.
1.0
Ability to use radius from annotation in NeuroML file for abstract cells - Many abstract cells in NeuroML2 have no 3D information, e.g. I&F, Izhikevich, specifically nothing saying what radius the should be when rendered in 3D. The default for this in OSB is to use 1um, as this has been useful for the C elegans model, e.g. [here](http://opensourcebrain.org/projects/celegans?explorer=https%253A%252F%252Fraw.githubusercontent.com%252Fopenworm%252FCElegansNeuroML%252Fmaster%252FCElegans%252FpythonScripts%252Fc302%252Fexamples%252Fc302_A_Muscles.nml#) since this appropriate for the radius for somas in the worm. Ideally the default should be 5um, but this could be overridden [in the `<population>` element](https://github.com/openworm/CElegansNeuroML/blob/master/CElegans/pythonScripts/c302/examples/c302_B_Muscles.nml#L59) in the neuroml file inside an `<annotation>` (similar to https://github.com/OpenSourceBrain/redmine/issues/244) allowing different networks to customise how they are displayed.
non_process
ability to use radius from annotation in neuroml file for abstract cells many abstract cells in have no information e g i f izhikevich specifically nothing saying what radius the should be when rendered in the default for this in osb is to use as this has been useful for the c elegans model e g since this appropriate for the radius for somas in the worm ideally the default should be but this could be overridden in the neuroml file inside an similar to allowing different networks to customise how they are displayed
0
143,474
13,065,174,897
IssuesEvent
2020-07-30 19:21:13
Bikossor/array-means
https://api.github.com/repos/Bikossor/array-means
closed
Split up the code examples
documentation
Split up the code examples _(Arithmetic mean, quadratic mean, etc.)_ in the `README.md` so it is more clear.
1.0
Split up the code examples - Split up the code examples _(Arithmetic mean, quadratic mean, etc.)_ in the `README.md` so it is more clear.
non_process
split up the code examples split up the code examples arithmetic mean quadratic mean etc in the readme md so it is more clear
0
13,439
15,882,111,966
IssuesEvent
2021-04-09 15:36:39
Sa-wol/ApartmentWebTemp
https://api.github.com/repos/Sa-wol/ApartmentWebTemp
opened
Set up Continuous Integration with Github Actions
dev process
Github actions should be used to build project on each commit
1.0
Set up Continuous Integration with Github Actions - Github actions should be used to build project on each commit
process
set up continuous integration with github actions github actions should be used to build project on each commit
1
360,465
10,693,164,524
IssuesEvent
2019-10-23 08:17:08
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
opened
z_sched_abort: sched_spinlock should be released before k_busy_wait
area: SMP bug priority: medium
**Describe the bug** In z_sched_abort of thread_abort.c ``` LOCKED(&sched_spinlock) { if (z_is_thread_prevented_from_running(thread)) { __ASSERT(!z_is_thread_queued(thread), ""); thread->base.thread_state |= _THREAD_DEAD; } else if (z_is_thread_queued(thread)) { _priq_run_remove(&_kernel.ready_q.runq, thread); z_mark_thread_as_not_queued(thread); thread->base.thread_state |= _THREAD_DEAD; } else { k_busy_wait(100); } } ``` the sched_spinlock should be released before k_busy_wait(100), if not released, other cores cannot get the sched_spinlock during k_busy_wait(100). This will worse the compete of locks, especially for cores >= 4
1.0
z_sched_abort: sched_spinlock should be released before k_busy_wait - **Describe the bug** In z_sched_abort of thread_abort.c ``` LOCKED(&sched_spinlock) { if (z_is_thread_prevented_from_running(thread)) { __ASSERT(!z_is_thread_queued(thread), ""); thread->base.thread_state |= _THREAD_DEAD; } else if (z_is_thread_queued(thread)) { _priq_run_remove(&_kernel.ready_q.runq, thread); z_mark_thread_as_not_queued(thread); thread->base.thread_state |= _THREAD_DEAD; } else { k_busy_wait(100); } } ``` the sched_spinlock should be released before k_busy_wait(100), if not released, other cores cannot get the sched_spinlock during k_busy_wait(100). This will worse the compete of locks, especially for cores >= 4
non_process
z sched abort sched spinlock should be released before k busy wait describe the bug in z sched abort of thread abort c locked sched spinlock if z is thread prevented from running thread assert z is thread queued thread thread base thread state thread dead else if z is thread queued thread priq run remove kernel ready q runq thread z mark thread as not queued thread thread base thread state thread dead else k busy wait the sched spinlock should be released before k busy wait if not released other cores cannot get the sched spinlock during k busy wait this will worse the compete of locks especially for cores
0
13,165
15,590,752,804
IssuesEvent
2021-03-18 09:44:13
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
opened
[FALSE-POSITIVE?] api.greenhouse.io
whitelisting process
Broken website: https://www.hackerrank.com/careers Note: Related to #171 (Creating a separate issue for ease of tracking per domain)
1.0
[FALSE-POSITIVE?] api.greenhouse.io - Broken website: https://www.hackerrank.com/careers Note: Related to #171 (Creating a separate issue for ease of tracking per domain)
process
api greenhouse io broken website note related to creating a separate issue for ease of tracking per domain
1
8,934
12,042,924,562
IssuesEvent
2020-04-14 11:29:22
pwittchen/ReactiveSensors
https://api.github.com/repos/pwittchen/ReactiveSensors
closed
release 0.4.0 [rx2, rx3]
release process
Release notes: - added new methods to `ReactiveSensorEvent` class: `int sensorId()`, `String sensorName()`, `float[] sensorValues()` Note: - for rx-3 I released 0.4.1 due to some issues with sonatype. I wasn't sure if the artifact for 0.4.0 was released or not
1.0
release 0.4.0 [rx2, rx3] - Release notes: - added new methods to `ReactiveSensorEvent` class: `int sensorId()`, `String sensorName()`, `float[] sensorValues()` Note: - for rx-3 I released 0.4.1 due to some issues with sonatype. I wasn't sure if the artifact for 0.4.0 was released or not
process
release release notes added new methods to reactivesensorevent class int sensorid string sensorname float sensorvalues note for rx i released due to some issues with sonatype i wasn t sure if the artifact for was released or not
1
313,776
23,491,362,704
IssuesEvent
2022-08-17 19:05:21
equinix/terraform-metal-openstack
https://api.github.com/repos/equinix/terraform-metal-openstack
closed
Missing TF_VAR documentation
bug documentation
README.md needs to include: export TF_VAR_metal_create_project=false OR the default needs to include this variable.
1.0
Missing TF_VAR documentation - README.md needs to include: export TF_VAR_metal_create_project=false OR the default needs to include this variable.
non_process
missing tf var documentation readme md needs to include export tf var metal create project false or the default needs to include this variable
0
14,669
17,787,553,304
IssuesEvent
2021-08-31 12:55:00
googleapis/python-crc32c
https://api.github.com/repos/googleapis/python-crc32c
reopened
Dependency Dashboard
type: process
This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Other Branches These updates are pending. To force PRs open, click the checkbox below. - [ ] <!-- other-branch=renovate/microsoft-setup-msbuild-1.x -->chore(deps): update microsoft/setup-msbuild action to v1.0.3 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Other Branches These updates are pending. To force PRs open, click the checkbox below. - [ ] <!-- other-branch=renovate/microsoft-setup-msbuild-1.x -->chore(deps): update microsoft/setup-msbuild action to v1.0.3 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue provides visibility into renovate updates and their statuses other branches these updates are pending to force prs open click the checkbox below chore deps update microsoft setup msbuild action to check this box to trigger a request for renovate to run again on this repository
1
57,685
11,777,987,352
IssuesEvent
2020-03-16 15:38:01
eventespresso/event-espresso-core
https://api.github.com/repos/eventespresso/event-espresso-core
closed
EntityList types, API and consumer components are inconsistent
EDTR v2 type:code-quality 👾
- `useTicketItem` and `useDatetimeItem` are being used inside a callback in `TableView.tsx` - `EntityList` uses `CardList` and `TableView` differently, while as they should behave in same manner with respect to what is passed to them. - Types defined in `/application/ui/layout/entityList/types.ts` are inconsistent - Many types in `/domain/eventEditor/interfaces/types.ts` are useless
1.0
EntityList types, API and consumer components are inconsistent - - `useTicketItem` and `useDatetimeItem` are being used inside a callback in `TableView.tsx` - `EntityList` uses `CardList` and `TableView` differently, while as they should behave in same manner with respect to what is passed to them. - Types defined in `/application/ui/layout/entityList/types.ts` are inconsistent - Many types in `/domain/eventEditor/interfaces/types.ts` are useless
non_process
entitylist types api and consumer components are inconsistent useticketitem and usedatetimeitem are being used inside a callback in tableview tsx entitylist uses cardlist and tableview differently while as they should behave in same manner with respect to what is passed to them types defined in application ui layout entitylist types ts are inconsistent many types in domain eventeditor interfaces types ts are useless
0
63,054
15,418,983,706
IssuesEvent
2021-03-05 09:32:54
Crocoblock/suggestions
https://api.github.com/repos/Crocoblock/suggestions
closed
JetWooBuilder - Custom category template is not really custom
JetWooBuilder Ready for release
**Describe the bug** In JetWooBuilder when create a custom shop template for a category, inside "Product loop" widget user cannot select custom archive template for that category. Instead default archive template is used. For ex. I created Shop default template--> with archive default list and grid templates, now I want to create Shop template for category1--> with custom archive for category 1, Shop template for category 2--> with custom archive for category 2. I cannot select custom archive template in Product Loop. I think that the main point for custom category templates is to be really custom with different look.
1.0
JetWooBuilder - Custom category template is not really custom - **Describe the bug** In JetWooBuilder when create a custom shop template for a category, inside "Product loop" widget user cannot select custom archive template for that category. Instead default archive template is used. For ex. I created Shop default template--> with archive default list and grid templates, now I want to create Shop template for category1--> with custom archive for category 1, Shop template for category 2--> with custom archive for category 2. I cannot select custom archive template in Product Loop. I think that the main point for custom category templates is to be really custom with different look.
non_process
jetwoobuilder custom category template is not really custom describe the bug in jetwoobuilder when create a custom shop template for a category inside product loop widget user cannot select custom archive template for that category instead default archive template is used for ex i created shop default template with archive default list and grid templates now i want to create shop template for with custom archive for category shop template for category with custom archive for category i cannot select custom archive template in product loop i think that the main point for custom category templates is to be really custom with different look
0
692,644
23,744,109,901
IssuesEvent
2022-08-31 14:39:12
phylum-dev/phylum-ci
https://api.github.com/repos/phylum-dev/phylum-ci
closed
Don't require serial processing of pre-commit hook
enhancement low priority python
## Overview **Is your feature request related to a problem? Please describe.** The current `phylum` pre-commit hook implementation requires the hook be run serially. The [`pre-commit` documentation for the `require_serial` option](https://pre-commit.com/index.html#hooks-require_serial) states that the hook: > ...will execute using a single process instead of in parallel. This is not really a problem as it does not impact hook execution performance since the `phylum` hook is expected to run against a single (preferably defined) dependency lockfile. That is, no more than one file will ever be processed by this hook. Even still, it is considered somewhat unusual to set `require_serial` and may confuse potential users who might otherwise assume the hook is less performant than it could be. **Describe the solution you'd like** The `phylum` pre-commit hook should not `require_serial` execution. **Describe alternatives you've considered** Continue with the existing hook implementation since it does work. However, the author of the `pre-commit` tool [has indicated that](https://github.com/pre-commit/pre-commit.com/pull/704#discussion_r957390361) setting `require_serial` may be a mis-configuration when multiprocessing is not otherwise provided by `phylum-ci`. ## Additional context The reason the current hook implementation is run serially is due to the logic for detecting a pre-commit environment and distinguishing it from other CI environments. If the `phylum-ci` script is provided extra arguments, it will assume that it is operating in a pre-commit environment because, as the [documentation for writing hooks states](https://pre-commit.com/index.html#arguments-pattern-in-hooks): > ...your hook should expect to receive the [`args`](https://pre-commit.com/index.html#config-args) value and then a list of staged files. The current implementation attempts to recognize the various ways a pre-commit environment can be entered and further distinguish them from simple user error, where unrecognized arguments were provided. It is this step that causes problems for parallel execution in a pre-commit environment. That is, with `require_serial` not set, it is possible that the list of staged files (or even all files when `pre-commit run --all-files` is used) is split up into chunks for parallel execution and all but one of those chunks will not contain the expected lockfile. Currently this results in a `SystemExit` error, which causes the hook to report a failure even when the chunk that does contain the lockfile passes. ## Acceptance criteria - [x] The `require_serial` flag in `.pre-commit-hooks.yaml` is removed - [x] The `phylum` pre-commit hook works when: - [x] All staged files are sent to the hook, whether the lockfile is included among them or not - [x] A filtered list of files are sent to the hook, whether the lockfile is included among them or not - [x] The lockfile is included in the extra arguments or explicitly specified - [x] Documentation/comments are updated
1.0
Don't require serial processing of pre-commit hook - ## Overview **Is your feature request related to a problem? Please describe.** The current `phylum` pre-commit hook implementation requires the hook be run serially. The [`pre-commit` documentation for the `require_serial` option](https://pre-commit.com/index.html#hooks-require_serial) states that the hook: > ...will execute using a single process instead of in parallel. This is not really a problem as it does not impact hook execution performance since the `phylum` hook is expected to run against a single (preferably defined) dependency lockfile. That is, no more than one file will ever be processed by this hook. Even still, it is considered somewhat unusual to set `require_serial` and may confuse potential users who might otherwise assume the hook is less performant than it could be. **Describe the solution you'd like** The `phylum` pre-commit hook should not `require_serial` execution. **Describe alternatives you've considered** Continue with the existing hook implementation since it does work. However, the author of the `pre-commit` tool [has indicated that](https://github.com/pre-commit/pre-commit.com/pull/704#discussion_r957390361) setting `require_serial` may be a mis-configuration when multiprocessing is not otherwise provided by `phylum-ci`. ## Additional context The reason the current hook implementation is run serially is due to the logic for detecting a pre-commit environment and distinguishing it from other CI environments. If the `phylum-ci` script is provided extra arguments, it will assume that it is operating in a pre-commit environment because, as the [documentation for writing hooks states](https://pre-commit.com/index.html#arguments-pattern-in-hooks): > ...your hook should expect to receive the [`args`](https://pre-commit.com/index.html#config-args) value and then a list of staged files. The current implementation attempts to recognize the various ways a pre-commit environment can be entered and further distinguish them from simple user error, where unrecognized arguments were provided. It is this step that causes problems for parallel execution in a pre-commit environment. That is, with `require_serial` not set, it is possible that the list of staged files (or even all files when `pre-commit run --all-files` is used) is split up into chunks for parallel execution and all but one of those chunks will not contain the expected lockfile. Currently this results in a `SystemExit` error, which causes the hook to report a failure even when the chunk that does contain the lockfile passes. ## Acceptance criteria - [x] The `require_serial` flag in `.pre-commit-hooks.yaml` is removed - [x] The `phylum` pre-commit hook works when: - [x] All staged files are sent to the hook, whether the lockfile is included among them or not - [x] A filtered list of files are sent to the hook, whether the lockfile is included among them or not - [x] The lockfile is included in the extra arguments or explicitly specified - [x] Documentation/comments are updated
non_process
don t require serial processing of pre commit hook overview is your feature request related to a problem please describe the current phylum pre commit hook implementation requires the hook be run serially the states that the hook will execute using a single process instead of in parallel this is not really a problem as it does not impact hook execution performance since the phylum hook is expected to run against a single preferably defined dependency lockfile that is no more than one file will ever be processed by this hook even still it is considered somewhat unusual to set require serial and may confuse potential users who might otherwise assume the hook is less performant than it could be describe the solution you d like the phylum pre commit hook should not require serial execution describe alternatives you ve considered continue with the existing hook implementation since it does work however the author of the pre commit tool setting require serial may be a mis configuration when multiprocessing is not otherwise provided by phylum ci additional context the reason the current hook implementation is run serially is due to the logic for detecting a pre commit environment and distinguishing it from other ci environments if the phylum ci script is provided extra arguments it will assume that it is operating in a pre commit environment because as the your hook should expect to receive the value and then a list of staged files the current implementation attempts to recognize the various ways a pre commit environment can be entered and further distinguish them from simple user error where unrecognized arguments were provided it is this step that causes problems for parallel execution in a pre commit environment that is with require serial not set it is possible that the list of staged files or even all files when pre commit run all files is used is split up into chunks for parallel execution and all but one of those chunks will not contain the expected lockfile currently this results in a systemexit error which causes the hook to report a failure even when the chunk that does contain the lockfile passes acceptance criteria the require serial flag in pre commit hooks yaml is removed the phylum pre commit hook works when all staged files are sent to the hook whether the lockfile is included among them or not a filtered list of files are sent to the hook whether the lockfile is included among them or not the lockfile is included in the extra arguments or explicitly specified documentation comments are updated
0
22,400
15,167,484,581
IssuesEvent
2021-02-12 17:51:55
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
dotnet-format breaks partial commits
area-infrastructure
I've been noticing some weirdness today. When I try to commit a subset of the files I've modified, or a subset of the lines within those files, sometimes my other modifications get committed too. This is causing some havoc with my coding workflow, making me think I was making all kinds of weird mistakes... :) Looking at `.githooks\pre-commit`, it doesn't appear to account for partial commits - it just fully adds all the files it touches. Am I missing something here that's supposed to make this work? I'm disabling it locally for now.
1.0
dotnet-format breaks partial commits - I've been noticing some weirdness today. When I try to commit a subset of the files I've modified, or a subset of the lines within those files, sometimes my other modifications get committed too. This is causing some havoc with my coding workflow, making me think I was making all kinds of weird mistakes... :) Looking at `.githooks\pre-commit`, it doesn't appear to account for partial commits - it just fully adds all the files it touches. Am I missing something here that's supposed to make this work? I'm disabling it locally for now.
non_process
dotnet format breaks partial commits i ve been noticing some weirdness today when i try to commit a subset of the files i ve modified or a subset of the lines within those files sometimes my other modifications get committed too this is causing some havoc with my coding workflow making me think i was making all kinds of weird mistakes looking at githooks pre commit it doesn t appear to account for partial commits it just fully adds all the files it touches am i missing something here that s supposed to make this work i m disabling it locally for now
0
98,258
16,361,478,316
IssuesEvent
2021-05-14 10:08:34
Galaxy-Software-Service/Maven_Pom_Demo
https://api.github.com/repos/Galaxy-Software-Service/Maven_Pom_Demo
opened
CVE-2016-3090 (High) detected in xwork-core-2.3.15.jar
security vulnerability
## CVE-2016-3090 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xwork-core-2.3.15.jar</b></p></summary> <p>Apache Struts 2</p> <p>Library home page: <a href="http://struts.apache.org/2.x/">http://struts.apache.org/2.x/</a></p> <p>Path to dependency file: Maven_Pom_Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.15/xwork-core-2.3.15.jar</p> <p> Dependency Hierarchy: - struts2-core-2.3.15.jar (Root Library) - :x: **xwork-core-2.3.15.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Galaxy-Software-Service/Maven_Pom_Demo/commit/3774e4b6d3e5f4ad3fa9f69e93f2b07cf1b66027">3774e4b6d3e5f4ad3fa9f69e93f2b07cf1b66027</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The TextParseUtil.translateVariables method in Apache Struts 2.x before 2.3.20 allows remote attackers to execute arbitrary code via a crafted OGNL expression with ANTLR tooling. <p>Publish Date: 2017-10-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-3090>CVE-2016-3090</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-3090">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-3090</a></p> <p>Release Date: 2017-10-30</p> <p>Fix Resolution: 2.3.20</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.struts.xwork","packageName":"xwork-core","packageVersion":"2.3.15","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.struts:struts2-core:2.3.15;org.apache.struts.xwork:xwork-core:2.3.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.20"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-3090","vulnerabilityDetails":"The TextParseUtil.translateVariables method in Apache Struts 2.x before 2.3.20 allows remote attackers to execute arbitrary code via a crafted OGNL expression with ANTLR tooling.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-3090","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-3090 (High) detected in xwork-core-2.3.15.jar - ## CVE-2016-3090 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xwork-core-2.3.15.jar</b></p></summary> <p>Apache Struts 2</p> <p>Library home page: <a href="http://struts.apache.org/2.x/">http://struts.apache.org/2.x/</a></p> <p>Path to dependency file: Maven_Pom_Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.15/xwork-core-2.3.15.jar</p> <p> Dependency Hierarchy: - struts2-core-2.3.15.jar (Root Library) - :x: **xwork-core-2.3.15.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Galaxy-Software-Service/Maven_Pom_Demo/commit/3774e4b6d3e5f4ad3fa9f69e93f2b07cf1b66027">3774e4b6d3e5f4ad3fa9f69e93f2b07cf1b66027</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The TextParseUtil.translateVariables method in Apache Struts 2.x before 2.3.20 allows remote attackers to execute arbitrary code via a crafted OGNL expression with ANTLR tooling. <p>Publish Date: 2017-10-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-3090>CVE-2016-3090</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-3090">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-3090</a></p> <p>Release Date: 2017-10-30</p> <p>Fix Resolution: 2.3.20</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.struts.xwork","packageName":"xwork-core","packageVersion":"2.3.15","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.struts:struts2-core:2.3.15;org.apache.struts.xwork:xwork-core:2.3.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.20"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-3090","vulnerabilityDetails":"The TextParseUtil.translateVariables method in Apache Struts 2.x before 2.3.20 allows remote attackers to execute arbitrary code via a crafted OGNL expression with ANTLR tooling.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-3090","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in xwork core jar cve high severity vulnerability vulnerable library xwork core jar apache struts library home page a href path to dependency file maven pom demo pom xml path to vulnerable library home wss scanner repository org apache struts xwork xwork core xwork core jar dependency hierarchy core jar root library x xwork core jar vulnerable library found in head commit a href found in base branch main vulnerability details the textparseutil translatevariables method in apache struts x before allows remote attackers to execute arbitrary code via a crafted ognl expression with antlr tooling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache struts core org apache struts xwork xwork core isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the textparseutil translatevariables method in apache struts x before allows remote attackers to execute arbitrary code via a crafted ognl expression with antlr tooling vulnerabilityurl
0
17,470
23,295,982,732
IssuesEvent
2022-08-06 15:23:25
Battle-s/battle-school-backend
https://api.github.com/repos/Battle-s/battle-school-backend
closed
[FEAT] Spring Security 세팅
feature :computer: processing :hourglass_flowing_sand:
## 설명 > 이슈에 대한 설명을 작성합니다. 담당자도 함께 작성하면 좋습니다. ## 체크사항 > 이슈를 close하기 위해 필요한 조건들을 체크박스로 나열합니다. - [x] config - [x] 인증 세팅 ## 참고자료 > 이슈를 해결하기 위해 필요한 참고자료가 있다면 추가합니다. ## 관련 논의 > 이슈에 대한 논의가 있었다면 논의 내용을 간략하게 추가합니다.
1.0
[FEAT] Spring Security 세팅 - ## 설명 > 이슈에 대한 설명을 작성합니다. 담당자도 함께 작성하면 좋습니다. ## 체크사항 > 이슈를 close하기 위해 필요한 조건들을 체크박스로 나열합니다. - [x] config - [x] 인증 세팅 ## 참고자료 > 이슈를 해결하기 위해 필요한 참고자료가 있다면 추가합니다. ## 관련 논의 > 이슈에 대한 논의가 있었다면 논의 내용을 간략하게 추가합니다.
process
spring security 세팅 설명 이슈에 대한 설명을 작성합니다 담당자도 함께 작성하면 좋습니다 체크사항 이슈를 close하기 위해 필요한 조건들을 체크박스로 나열합니다 config 인증 세팅 참고자료 이슈를 해결하기 위해 필요한 참고자료가 있다면 추가합니다 관련 논의 이슈에 대한 논의가 있었다면 논의 내용을 간략하게 추가합니다
1
419,655
28,149,721,558
IssuesEvent
2023-04-02 22:12:50
aansbacher/debate
https://api.github.com/repos/aansbacher/debate
closed
תרשים זרימת משתמש
documentation
התחברות/הרשמה בחירת נושא - מבין רשימת נושאים משתתף/צופה/שופט. רמת ההבנה שלך בנושא - מומחה, לא פראייר, באתי ללמוד. תזמן/ כנס עכשיו לזירת וידאו, צ'אט
1.0
תרשים זרימת משתמש - התחברות/הרשמה בחירת נושא - מבין רשימת נושאים משתתף/צופה/שופט. רמת ההבנה שלך בנושא - מומחה, לא פראייר, באתי ללמוד. תזמן/ כנס עכשיו לזירת וידאו, צ'אט
non_process
תרשים זרימת משתמש התחברות הרשמה בחירת נושא מבין רשימת נושאים משתתף צופה שופט רמת ההבנה שלך בנושא מומחה לא פראייר באתי ללמוד תזמן כנס עכשיו לזירת וידאו צ אט
0
469,341
13,506,854,298
IssuesEvent
2020-09-14 04:20:42
Kedyn/fusliez-notes
https://api.github.com/repos/Kedyn/fusliez-notes
closed
Project structure
Priority: Critical Status: In Progress Type: Maintenance Type: Question
As mention on issue #5 we might want to follow the following structure: ``` ./ComponentName/ The folder containing the component ComponentName.styles.ts The file containing the components styles ComponentName.tsx The actual component index.ts File that exports the component so imports can just be /components/ComponentName/ instead of /component/ComponentName/ComponentName ``` There is a VSCode extension that generates this files called **VSCode React Component Generator**. I also have a few snippets I could put in the repo so we can all follow the same/similar format. - [x] Main content (score /players section) - [x] Controls content ( Score buttons, reset button / notes) - [x] Map content
1.0
Project structure - As mention on issue #5 we might want to follow the following structure: ``` ./ComponentName/ The folder containing the component ComponentName.styles.ts The file containing the components styles ComponentName.tsx The actual component index.ts File that exports the component so imports can just be /components/ComponentName/ instead of /component/ComponentName/ComponentName ``` There is a VSCode extension that generates this files called **VSCode React Component Generator**. I also have a few snippets I could put in the repo so we can all follow the same/similar format. - [x] Main content (score /players section) - [x] Controls content ( Score buttons, reset button / notes) - [x] Map content
non_process
project structure as mention on issue we might want to follow the following structure componentname the folder containing the component componentname styles ts the file containing the components styles componentname tsx the actual component index ts file that exports the component so imports can just be components componentname instead of component componentname componentname there is a vscode extension that generates this files called vscode react component generator i also have a few snippets i could put in the repo so we can all follow the same similar format main content score players section controls content score buttons reset button notes map content
0
235,844
7,743,268,875
IssuesEvent
2018-05-29 12:19:22
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0000650: rename functions in Admin Controllers
Admin low priority
**Reported by pschuele on 9 Dec 2008 12:24** rename the functions: - getXXXs -&gt; search - getXXX -&gt; get - addXXX -&gt; create - updateXXX -&gt; update - deleteXXXs -&gt; delete
1.0
0000650: rename functions in Admin Controllers - **Reported by pschuele on 9 Dec 2008 12:24** rename the functions: - getXXXs -&gt; search - getXXX -&gt; get - addXXX -&gt; create - updateXXX -&gt; update - deleteXXXs -&gt; delete
non_process
rename functions in admin controllers reported by pschuele on dec rename the functions getxxxs gt search getxxx gt get addxxx gt create updatexxx gt update deletexxxs gt delete
0
2,995
5,970,154,425
IssuesEvent
2017-05-30 21:54:45
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Cannot kill child process in Ubuntu 14.04 x64 (on tutum's docker build)
child_process
- **Version**: _5.7_ - **Platform**: _Ubuntu 14.04.4 x64_ - **Subsystem**: _docker build + npm test command for build purposes testing a restify api with mocha and supertest_ I have a simple restify api tested with mocha and supertest. When I run the tests locally on my mac they run perfectly by: 1. starting a `node server.js` child process 2. running all the tests 3. and killing the process running `node server.js` Since version 5.7 I cannot kill the process `node server.js` anymore on Ubuntu. ## In server.js I have a gracefullyExit function to test which signals is being called ``` javascript var gracefullyExit = function(signal) { shuttingdown = true; log.info('Gracefully Exit Called (' + signal + ')'); server.close(function() { process.exit(0); }); }; process.on('uncaughtException', function(error) { log.error('HiddenException:', error); }); process.on('SIGTERM', function() { gracefullyExit('SIGTERM'); }); process.on('SIGKILL', function() { gracefullyExit('SIGKILL'); }); process.on('SIGINT', function() { gracefullyExit('SIGINT'); }); process.on('SIGABRT', function() { gracefullyExit('SIGABRT'); }); process.on('SIGHUP', function() { gracefullyExit('SIGHUP'); }); process.on('SIGQUIT', function() { gracefullyExit('SIGQUIT'); }); ``` In my mocha's after() I have tried all of these: ``` javascript after(function(done) { this.timeout(5000); process.env.NODE_ENV = environment; if(os.type() === 'Linux') { spawn('kill -9 ' + server.pid); } process.kill(server.pid); server.stdout.on('data', function(data) { if (SERVER_LOGGING) { console.log(data); } if (data.indexOf('Analyze server closing') !== -1) { console.log(''); console.log(''); console.log(data); done(); } }); }); ``` And This before that last one: ``` javascript after(function(done) { this.timeout(5000); process.env.NODE_ENV = environment; process.kill(server.pid, 'SIGTERM'); server.stdout.on('data', function(data) { if (SERVER_LOGGING) { console.log(data); } if (data.indexOf('Gracefully Exit Called') !== -1) { console.log(''); console.log(''); console.log(data); done(); } }); }); ``` All with all different SIGNALS (SIGINT, SIGTERM, SIGKILL, SIGQUIT, etc) ## Nothing works on UBUNTU, everything works on MAC. **RESULTING IN a TIMEOUT on Ubuntu** ``` bash 1) "after all" hook 163 passing (9s) 1 failing 1) Test "after all" hook: Error: timeout of 5000ms exceeded. Ensure the done() callback is being called in this test. ``` How can I spawn and kill a nodejs server from a nodejs mocha test? If I can't kill the `node server.js` I cannot launch another test suite. And if I disregard the actual closing of the server and termination of the child process running `node server.js` I end up getting the error EADDRINUSE etc because the process is still alive. ### THIS USED TO WORK SOMEHOW, not anymore
1.0
Cannot kill child process in Ubuntu 14.04 x64 (on tutum's docker build) - - **Version**: _5.7_ - **Platform**: _Ubuntu 14.04.4 x64_ - **Subsystem**: _docker build + npm test command for build purposes testing a restify api with mocha and supertest_ I have a simple restify api tested with mocha and supertest. When I run the tests locally on my mac they run perfectly by: 1. starting a `node server.js` child process 2. running all the tests 3. and killing the process running `node server.js` Since version 5.7 I cannot kill the process `node server.js` anymore on Ubuntu. ## In server.js I have a gracefullyExit function to test which signals is being called ``` javascript var gracefullyExit = function(signal) { shuttingdown = true; log.info('Gracefully Exit Called (' + signal + ')'); server.close(function() { process.exit(0); }); }; process.on('uncaughtException', function(error) { log.error('HiddenException:', error); }); process.on('SIGTERM', function() { gracefullyExit('SIGTERM'); }); process.on('SIGKILL', function() { gracefullyExit('SIGKILL'); }); process.on('SIGINT', function() { gracefullyExit('SIGINT'); }); process.on('SIGABRT', function() { gracefullyExit('SIGABRT'); }); process.on('SIGHUP', function() { gracefullyExit('SIGHUP'); }); process.on('SIGQUIT', function() { gracefullyExit('SIGQUIT'); }); ``` In my mocha's after() I have tried all of these: ``` javascript after(function(done) { this.timeout(5000); process.env.NODE_ENV = environment; if(os.type() === 'Linux') { spawn('kill -9 ' + server.pid); } process.kill(server.pid); server.stdout.on('data', function(data) { if (SERVER_LOGGING) { console.log(data); } if (data.indexOf('Analyze server closing') !== -1) { console.log(''); console.log(''); console.log(data); done(); } }); }); ``` And This before that last one: ``` javascript after(function(done) { this.timeout(5000); process.env.NODE_ENV = environment; process.kill(server.pid, 'SIGTERM'); server.stdout.on('data', function(data) { if (SERVER_LOGGING) { console.log(data); } if (data.indexOf('Gracefully Exit Called') !== -1) { console.log(''); console.log(''); console.log(data); done(); } }); }); ``` All with all different SIGNALS (SIGINT, SIGTERM, SIGKILL, SIGQUIT, etc) ## Nothing works on UBUNTU, everything works on MAC. **RESULTING IN a TIMEOUT on Ubuntu** ``` bash 1) "after all" hook 163 passing (9s) 1 failing 1) Test "after all" hook: Error: timeout of 5000ms exceeded. Ensure the done() callback is being called in this test. ``` How can I spawn and kill a nodejs server from a nodejs mocha test? If I can't kill the `node server.js` I cannot launch another test suite. And if I disregard the actual closing of the server and termination of the child process running `node server.js` I end up getting the error EADDRINUSE etc because the process is still alive. ### THIS USED TO WORK SOMEHOW, not anymore
process
cannot kill child process in ubuntu on tutum s docker build version platform ubuntu subsystem docker build npm test command for build purposes testing a restify api with mocha and supertest i have a simple restify api tested with mocha and supertest when i run the tests locally on my mac they run perfectly by starting a node server js child process running all the tests and killing the process running node server js since version i cannot kill the process node server js anymore on ubuntu in server js i have a gracefullyexit function to test which signals is being called javascript var gracefullyexit function signal shuttingdown true log info gracefully exit called signal server close function process exit process on uncaughtexception function error log error hiddenexception error process on sigterm function gracefullyexit sigterm process on sigkill function gracefullyexit sigkill process on sigint function gracefullyexit sigint process on sigabrt function gracefullyexit sigabrt process on sighup function gracefullyexit sighup process on sigquit function gracefullyexit sigquit in my mocha s after i have tried all of these javascript after function done this timeout process env node env environment if os type linux spawn kill server pid process kill server pid server stdout on data function data if server logging console log data if data indexof analyze server closing console log console log console log data done and this before that last one javascript after function done this timeout process env node env environment process kill server pid sigterm server stdout on data function data if server logging console log data if data indexof gracefully exit called console log console log console log data done all with all different signals sigint sigterm sigkill sigquit etc nothing works on ubuntu everything works on mac resulting in a timeout on ubuntu bash after all hook passing failing test after all hook error timeout of exceeded ensure the done callback is being called in this test how can i spawn and kill a nodejs server from a nodejs mocha test if i can t kill the node server js i cannot launch another test suite and if i disregard the actual closing of the server and termination of the child process running node server js i end up getting the error eaddrinuse etc because the process is still alive this used to work somehow not anymore
1
7,189
9,475,373,707
IssuesEvent
2019-04-19 11:00:41
ballerina-platform/ballerina-spec
https://api.github.com/repos/ballerina-platform/ballerina-spec
closed
Improve error constructor syntax
enhancement incompatible lang status/agreed
**Problem:** Error constructors are problematic in that users cannot take advantage of an error type definition to avoid repeating the error reason string. Code sample that shows issue: type IOError error<"ioerror", record { string filename?; }>; function foo(string s) returns IOError? { return error("ioerror", {filename: s}); } **Solution:** Improve error constructor syntax by - specifying the fields as named arguments, and - making error reason defaultable. The error constructor in the example would then become ``` return error(filename = s); ``` The fields would be defaultable from the record type as with the mapping constructor. This also needs to be done for binding patterns. This is a prerequisite for #2.
True
Improve error constructor syntax - **Problem:** Error constructors are problematic in that users cannot take advantage of an error type definition to avoid repeating the error reason string. Code sample that shows issue: type IOError error<"ioerror", record { string filename?; }>; function foo(string s) returns IOError? { return error("ioerror", {filename: s}); } **Solution:** Improve error constructor syntax by - specifying the fields as named arguments, and - making error reason defaultable. The error constructor in the example would then become ``` return error(filename = s); ``` The fields would be defaultable from the record type as with the mapping constructor. This also needs to be done for binding patterns. This is a prerequisite for #2.
non_process
improve error constructor syntax problem error constructors are problematic in that users cannot take advantage of an error type definition to avoid repeating the error reason string code sample that shows issue type ioerror error function foo string s returns ioerror return error ioerror filename s solution improve error constructor syntax by specifying the fields as named arguments and making error reason defaultable the error constructor in the example would then become return error filename s the fields would be defaultable from the record type as with the mapping constructor this also needs to be done for binding patterns this is a prerequisite for
0
7,916
11,096,880,560
IssuesEvent
2019-12-16 12:09:44
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Diagnostics.AspNet: Add integration tests against an IIS hosted application (Web Api and MVC)
api: clouderrorreporting type: process
There are differences between self hosted and IIS hosted ASP.Net applications that impact the Diagnostics library ( see #2292 and #2294). We are adding integration tests against self hosted applications but we should also add test applications against IIS hosted applications, since this is probably the most common use case for ASP.Net development. (We can probably set up an IIS Express on the jenkins machine an have a simple ASP.Net application hosted there to test against). This is way less important for ASP.Net Core because ASP.Net Core is supposed to run the same "everywhere", so it is more OK if we assume the same behavior from ASP.Net Core applications regardless of how they are being hosted. And we already have integrations tests for ASP.Net Core.
1.0
Diagnostics.AspNet: Add integration tests against an IIS hosted application (Web Api and MVC) - There are differences between self hosted and IIS hosted ASP.Net applications that impact the Diagnostics library ( see #2292 and #2294). We are adding integration tests against self hosted applications but we should also add test applications against IIS hosted applications, since this is probably the most common use case for ASP.Net development. (We can probably set up an IIS Express on the jenkins machine an have a simple ASP.Net application hosted there to test against). This is way less important for ASP.Net Core because ASP.Net Core is supposed to run the same "everywhere", so it is more OK if we assume the same behavior from ASP.Net Core applications regardless of how they are being hosted. And we already have integrations tests for ASP.Net Core.
process
diagnostics aspnet add integration tests against an iis hosted application web api and mvc there are differences between self hosted and iis hosted asp net applications that impact the diagnostics library see and we are adding integration tests against self hosted applications but we should also add test applications against iis hosted applications since this is probably the most common use case for asp net development we can probably set up an iis express on the jenkins machine an have a simple asp net application hosted there to test against this is way less important for asp net core because asp net core is supposed to run the same everywhere so it is more ok if we assume the same behavior from asp net core applications regardless of how they are being hosted and we already have integrations tests for asp net core
1
9,924
12,963,469,714
IssuesEvent
2020-07-20 18:50:31
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
reopened
misleading categorization - azure devops server 2019 does not support parameters
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
The selector in the upper left corner is misleading users - it is possible to select server 2019 but according to https://docs.microsoft.com/en-us/azure/devops/release-notes/features-timeline it will only be supported in server 2020 --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 790318bb-8220-3241-4ca7-73351074492f * Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86 * Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script) * Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
misleading categorization - azure devops server 2019 does not support parameters - The selector in the upper left corner is misleading users - it is possible to select server 2019 but according to https://docs.microsoft.com/en-us/azure/devops/release-notes/features-timeline it will only be supported in server 2020 --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 790318bb-8220-3241-4ca7-73351074492f * Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86 * Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script) * Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
misleading categorization azure devops server does not support parameters the selector in the upper left corner is misleading users it is possible to select server but according to it will only be supported in server document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
73,755
7,354,388,040
IssuesEvent
2018-03-09 06:34:50
pypa/warehouse
https://api.github.com/repos/pypa/warehouse
opened
Improve links in management UI
HTML good first issue user testing
## Release page on `/manage/project/<project>/releases/`, we link to `/manage/project/<project>/releases/<release>` in two places: ![screenshot from 2018-03-09 06-25-47](https://user-images.githubusercontent.com/3323703/37193709-f319f740-2362-11e8-84b1-dcc7231c3fa3.png) ![screenshot from 2018-03-09 06-25-56](https://user-images.githubusercontent.com/3323703/37193712-f358ce48-2362-11e8-825f-56e3bb36de16.png) We should: 1. Change the 'edit' dropdown to say 'Manage' 2. Add a "Manage version" title to the version number link ## Release detail page On the release detail page: `/manage/project/<project>/releases/<release>`: ![screenshot from 2018-03-09 06-25-38](https://user-images.githubusercontent.com/3323703/37193710-f336228a-2362-11e8-818f-2dccc4b47d27.png) We should: 1. Add a "Download file" title to the download link 2. Restructure the HTML so that the link underline does not extend beyond the text (currently, the space between the text and the file size is underlined) --- **Good First Issue**: This issue is good for first time contributors. If you've already contributed to Warehouse, please work on [another issue without this label](https://github.com/pypa/warehouse/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aopen+-label%3A%22good+first+issue%22) instead. If there is not a corresponding pull request for this issue, it is up for grabs. For directions for getting set up, see our [Getting Started Guide](https://warehouse.pypa.io/development/getting-started/). If you are working on this issue and have questions, please feel free to ask them here, [`#pypa-dev` on Freenode](https://webchat.freenode.net/?channels=%23pypa-dev), or the [pypa-dev mailing list](https://groups.google.com/forum/#!forum/pypa-dev).
1.0
Improve links in management UI - ## Release page on `/manage/project/<project>/releases/`, we link to `/manage/project/<project>/releases/<release>` in two places: ![screenshot from 2018-03-09 06-25-47](https://user-images.githubusercontent.com/3323703/37193709-f319f740-2362-11e8-84b1-dcc7231c3fa3.png) ![screenshot from 2018-03-09 06-25-56](https://user-images.githubusercontent.com/3323703/37193712-f358ce48-2362-11e8-825f-56e3bb36de16.png) We should: 1. Change the 'edit' dropdown to say 'Manage' 2. Add a "Manage version" title to the version number link ## Release detail page On the release detail page: `/manage/project/<project>/releases/<release>`: ![screenshot from 2018-03-09 06-25-38](https://user-images.githubusercontent.com/3323703/37193710-f336228a-2362-11e8-818f-2dccc4b47d27.png) We should: 1. Add a "Download file" title to the download link 2. Restructure the HTML so that the link underline does not extend beyond the text (currently, the space between the text and the file size is underlined) --- **Good First Issue**: This issue is good for first time contributors. If you've already contributed to Warehouse, please work on [another issue without this label](https://github.com/pypa/warehouse/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aopen+-label%3A%22good+first+issue%22) instead. If there is not a corresponding pull request for this issue, it is up for grabs. For directions for getting set up, see our [Getting Started Guide](https://warehouse.pypa.io/development/getting-started/). If you are working on this issue and have questions, please feel free to ask them here, [`#pypa-dev` on Freenode](https://webchat.freenode.net/?channels=%23pypa-dev), or the [pypa-dev mailing list](https://groups.google.com/forum/#!forum/pypa-dev).
non_process
improve links in management ui release page on manage project releases we link to manage project releases in two places we should change the edit dropdown to say manage add a manage version title to the version number link release detail page on the release detail page manage project releases we should add a download file title to the download link restructure the html so that the link underline does not extend beyond the text currently the space between the text and the file size is underlined good first issue this issue is good for first time contributors if you ve already contributed to warehouse please work on instead if there is not a corresponding pull request for this issue it is up for grabs for directions for getting set up see our if you are working on this issue and have questions please feel free to ask them here or the
0
159,304
12,474,280,884
IssuesEvent
2020-05-29 09:22:13
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
Using xml., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552
bandit bug static analysis testing triage
```yaml { "id": 1, "title": "Using xml., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Using xml.dom.minidom.parseString to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.dom.minidom.parseString with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called at /opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552 See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-29 (09:21)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-29 (09:21)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "/opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
Using xml., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552 - ```yaml { "id": 1, "title": "Using xml., /opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Using xml.dom.minidom.parseString to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.dom.minidom.parseString with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called at /opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552 See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-29 (09:21)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-29 (09:21)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "/opt/ros_noetic_ws/src/ros/roslib/src/roslib/manifestlib.py:552", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_process
using xml opt ros noetic ws src ros roslib src roslib manifestlib py yaml id title using xml opt ros noetic ws src ros roslib src roslib manifestlib py type bug description high confidence of medium severity bug using xml dom minidom parsestring to parse untrusted xml data is known to be vulnerable to xml attacks replace xml dom minidom parsestring with its defusedxml equivalent function or make sure defusedxml defuse stdlib is called at opt ros noetic ws src ros roslib src roslib manifestlib py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace opt ros noetic ws src ros roslib src roslib manifestlib py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
36
2,506,194,686
IssuesEvent
2015-01-12 07:44:12
sysown/proxysql-0.2
https://api.github.com/repos/sysown/proxysql-0.2
closed
Define criteria of when threads need to update their QP statistics
ADMIN development QUERY PROCESSOR
There is no defined criteria of when a thread should update its QP statistics
1.0
Define criteria of when threads need to update their QP statistics - There is no defined criteria of when a thread should update its QP statistics
process
define criteria of when threads need to update their qp statistics there is no defined criteria of when a thread should update its qp statistics
1
5,265
8,057,651,132
IssuesEvent
2018-08-02 15:57:34
GoogleCloudPlatform/google-cloud-python
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
closed
BigQuery: 'Client.update_table' systest error, 412 PreconditionFailed
api: bigquery backend flaky testing type: process
Second error in [this CI run](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7126): ```python ____________________ TestBigQuery.test_update_table_schema _____________________ self = <tests.system.TestBigQuery testMethod=test_update_table_schema> def test_update_table_schema(self): dataset = self.temp_dataset(_make_dataset_id('update_table')) TABLE_NAME = 'test_table' table_arg = Table(dataset.table(TABLE_NAME), schema=SCHEMA) self.assertFalse(_table_exists(table_arg)) table = retry_403(Config.CLIENT.create_table)(table_arg) self.to_delete.insert(0, table) self.assertTrue(_table_exists(table)) voter = bigquery.SchemaField('voter', 'BOOLEAN', mode='NULLABLE') schema = table.schema schema.append(voter) table.schema = schema > updated_table = Config.CLIENT.update_table(table, ['schema']) ,,, if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E PreconditionFailed: 412 PATCH https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/update_table_7126_1531263328/tables/test_table: Precondition Failed ../.nox/sys-2-7/lib/python2.7/site-packages/google/cloud/_http.py:293: PreconditionFailed ```
1.0
BigQuery: 'Client.update_table' systest error, 412 PreconditionFailed - Second error in [this CI run](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7126): ```python ____________________ TestBigQuery.test_update_table_schema _____________________ self = <tests.system.TestBigQuery testMethod=test_update_table_schema> def test_update_table_schema(self): dataset = self.temp_dataset(_make_dataset_id('update_table')) TABLE_NAME = 'test_table' table_arg = Table(dataset.table(TABLE_NAME), schema=SCHEMA) self.assertFalse(_table_exists(table_arg)) table = retry_403(Config.CLIENT.create_table)(table_arg) self.to_delete.insert(0, table) self.assertTrue(_table_exists(table)) voter = bigquery.SchemaField('voter', 'BOOLEAN', mode='NULLABLE') schema = table.schema schema.append(voter) table.schema = schema > updated_table = Config.CLIENT.update_table(table, ['schema']) ,,, if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E PreconditionFailed: 412 PATCH https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/update_table_7126_1531263328/tables/test_table: Precondition Failed ../.nox/sys-2-7/lib/python2.7/site-packages/google/cloud/_http.py:293: PreconditionFailed ```
process
bigquery client update table systest error preconditionfailed second error in python testbigquery test update table schema self def test update table schema self dataset self temp dataset make dataset id update table table name test table table arg table dataset table table name schema schema self assertfalse table exists table arg table retry config client create table table arg self to delete insert table self asserttrue table exists table voter bigquery schemafield voter boolean mode nullable schema table schema schema append voter table schema schema updated table config client update table table if not response status code raise exceptions from http response response e preconditionfailed patch precondition failed nox sys lib site packages google cloud http py preconditionfailed
1
40,055
8,722,057,211
IssuesEvent
2018-12-09 07:59:56
publiclab/plots2
https://api.github.com/repos/publiclab/plots2
closed
Enhance readability of text in small screens
add-code-links gci-candidate small
```We are preparing to participate in Google Code-in, and have reserved this issue for participants in GCI - but we'd love to have your help with another one! Please check out https://code.publiclab.org to see more.``` Hi, this is a [first-timers-only issue](https://publiclab.github.io/community-toolbox/#r=all). This means we've worked to make it more legible to folks who either **haven't contributed to our codebase before, or even folks who haven't contributed to open source before**. If that's you, we're interested in helping you take the first step and can answer questions and help you out as you do. Note that we're especially interested in contributions from people from groups underrepresented in free and open source software! We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝 If you have contributed before, **consider leaving this one for someone new**, and looking through our general [help wanted](https://github.com/publiclab/plots2/labels/help-wanted) issues. Thanks! ### 🤔 What you will need to know. Nothing. This issue is meant to welcome you to Open Source :) We are happy to walk you through the process. ### 📋 Step by Step - [ ] 🙋 **Claim this issue**: Comment below. If someone else has claimed it, ask if they've opened a pull request already and if they're stuck -- maybe you can help them solve a problem or move it along! - [ ] 📝 **Update** the text color of **username**,**password**,**remember me** to white so that it is visible in mobile browser ![color_problem](https://user-images.githubusercontent.com/24489162/46003987-6defde80-c0cf-11e8-8f97-c556527c24b8.jpeg) The above text is displayed on the dashboard whenever a person signs up at Public Labs for teh first time - [ ] 💾 **Commit** your changes - [ ] 🔀 **Start a Pull Request**. There are two ways how you can start a pull request: 1. If you are familiar with the terminal or would like to learn it, [here is a great tutorial](https://egghead.io/series/how-to-contribute-to-an-open-source-project-on-github) on how to send a pull request using the terminal. 2. You can also [edit files directly in your browser](https://help.github.com/articles/editing-files-in-your-repository/) and open a pull request from there. - [ ] 🏁 **Done** Ask in comments for a review :) ### 🤔❓ Questions? Leave a comment below! ### Is someone else already working on this? We encourage you to link to this issue by mentioning the issue # in your pull request, so we can see if someone's already started on it. **If someone seem stuck, offer them some help!** Otherwise, [take a look at some other issues you can help with](https://publiclab.github.io/community-toolbox/#r=all). Thanks! (This issue was created by [First-Timers-Bot](https://github.com/hoodiehq/first-timers-bot).)
1.0
Enhance readability of text in small screens - ```We are preparing to participate in Google Code-in, and have reserved this issue for participants in GCI - but we'd love to have your help with another one! Please check out https://code.publiclab.org to see more.``` Hi, this is a [first-timers-only issue](https://publiclab.github.io/community-toolbox/#r=all). This means we've worked to make it more legible to folks who either **haven't contributed to our codebase before, or even folks who haven't contributed to open source before**. If that's you, we're interested in helping you take the first step and can answer questions and help you out as you do. Note that we're especially interested in contributions from people from groups underrepresented in free and open source software! We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝 If you have contributed before, **consider leaving this one for someone new**, and looking through our general [help wanted](https://github.com/publiclab/plots2/labels/help-wanted) issues. Thanks! ### 🤔 What you will need to know. Nothing. This issue is meant to welcome you to Open Source :) We are happy to walk you through the process. ### 📋 Step by Step - [ ] 🙋 **Claim this issue**: Comment below. If someone else has claimed it, ask if they've opened a pull request already and if they're stuck -- maybe you can help them solve a problem or move it along! - [ ] 📝 **Update** the text color of **username**,**password**,**remember me** to white so that it is visible in mobile browser ![color_problem](https://user-images.githubusercontent.com/24489162/46003987-6defde80-c0cf-11e8-8f97-c556527c24b8.jpeg) The above text is displayed on the dashboard whenever a person signs up at Public Labs for teh first time - [ ] 💾 **Commit** your changes - [ ] 🔀 **Start a Pull Request**. There are two ways how you can start a pull request: 1. If you are familiar with the terminal or would like to learn it, [here is a great tutorial](https://egghead.io/series/how-to-contribute-to-an-open-source-project-on-github) on how to send a pull request using the terminal. 2. You can also [edit files directly in your browser](https://help.github.com/articles/editing-files-in-your-repository/) and open a pull request from there. - [ ] 🏁 **Done** Ask in comments for a review :) ### 🤔❓ Questions? Leave a comment below! ### Is someone else already working on this? We encourage you to link to this issue by mentioning the issue # in your pull request, so we can see if someone's already started on it. **If someone seem stuck, offer them some help!** Otherwise, [take a look at some other issues you can help with](https://publiclab.github.io/community-toolbox/#r=all). Thanks! (This issue was created by [First-Timers-Bot](https://github.com/hoodiehq/first-timers-bot).)
non_process
enhance readability of text in small screens we are preparing to participate in google code in and have reserved this issue for participants in gci but we d love to have your help with another one please check out to see more hi this is a this means we ve worked to make it more legible to folks who either haven t contributed to our codebase before or even folks who haven t contributed to open source before if that s you we re interested in helping you take the first step and can answer questions and help you out as you do note that we re especially interested in contributions from people from groups underrepresented in free and open source software we know that the process of creating a pull request is the biggest barrier for new contributors this issue is for you 💝 if you have contributed before consider leaving this one for someone new and looking through our general issues thanks 🤔 what you will need to know nothing this issue is meant to welcome you to open source we are happy to walk you through the process 📋 step by step 🙋 claim this issue comment below if someone else has claimed it ask if they ve opened a pull request already and if they re stuck maybe you can help them solve a problem or move it along 📝 update the text color of username password remember me to white so that it is visible in mobile browser the above text is displayed on the dashboard whenever a person signs up at public labs for teh first time 💾 commit your changes 🔀 start a pull request there are two ways how you can start a pull request if you are familiar with the terminal or would like to learn it on how to send a pull request using the terminal you can also and open a pull request from there 🏁 done ask in comments for a review 🤔❓ questions leave a comment below is someone else already working on this we encourage you to link to this issue by mentioning the issue in your pull request so we can see if someone s already started on it if someone seem stuck offer them some help otherwise thanks this issue was created by
0
18,027
24,035,741,579
IssuesEvent
2022-09-15 18:59:31
magland/spikesortingview
https://api.github.com/repos/magland/spikesortingview
closed
option for display min/max on y axis
in process
This should be an option rather than hard-coded one way or the other _Originally posted by @jsoules in https://github.com/magland/spikesortingview/pull/120#discussion_r969828466_
1.0
option for display min/max on y axis - This should be an option rather than hard-coded one way or the other _Originally posted by @jsoules in https://github.com/magland/spikesortingview/pull/120#discussion_r969828466_
process
option for display min max on y axis this should be an option rather than hard coded one way or the other originally posted by jsoules in
1
149,149
19,566,178,343
IssuesEvent
2022-01-04 00:54:28
ghc-dev/Cole-Pineda
https://api.github.com/repos/ghc-dev/Cole-Pineda
opened
CVE-2020-26137 (Medium) detected in urllib3-1.24.3-py2.py3-none-any.whl
security vulnerability
## CVE-2020-26137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.24.3-py2.py3-none-any.whl</b></p></summary> <p>HTTP library with thread-safe connection pooling, file post, and more.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/01/11/525b02e4acc0c747de8b6ccdab376331597c569c42ea66ab0a1dbd36eca2/urllib3-1.24.3-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/01/11/525b02e4acc0c747de8b6ccdab376331597c569c42ea66ab0a1dbd36eca2/urllib3-1.24.3-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - requests-2.20.0-py2.py3-none-any.whl (Root Library) - :x: **urllib3-1.24.3-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Cole-Pineda/commit/b066a30ffdb45d169672eff2d276def5b807539f">b066a30ffdb45d169672eff2d276def5b807539f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> urllib3 before 1.25.9 allows CRLF injection if the attacker controls the HTTP request method, as demonstrated by inserting CR and LF control characters in the first argument of putrequest(). NOTE: this is similar to CVE-2020-26116. <p>Publish Date: 2020-09-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26137>CVE-2020-26137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26137">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26137</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: 1.25.9</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"urllib3","packageVersion":"1.24.3","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":true,"dependencyTree":"requests:2.20.0;urllib3:1.24.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.25.9","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-26137","vulnerabilityDetails":"urllib3 before 1.25.9 allows CRLF injection if the attacker controls the HTTP request method, as demonstrated by inserting CR and LF control characters in the first argument of putrequest(). NOTE: this is similar to CVE-2020-26116.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26137","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-26137 (Medium) detected in urllib3-1.24.3-py2.py3-none-any.whl - ## CVE-2020-26137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.24.3-py2.py3-none-any.whl</b></p></summary> <p>HTTP library with thread-safe connection pooling, file post, and more.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/01/11/525b02e4acc0c747de8b6ccdab376331597c569c42ea66ab0a1dbd36eca2/urllib3-1.24.3-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/01/11/525b02e4acc0c747de8b6ccdab376331597c569c42ea66ab0a1dbd36eca2/urllib3-1.24.3-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - requests-2.20.0-py2.py3-none-any.whl (Root Library) - :x: **urllib3-1.24.3-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Cole-Pineda/commit/b066a30ffdb45d169672eff2d276def5b807539f">b066a30ffdb45d169672eff2d276def5b807539f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> urllib3 before 1.25.9 allows CRLF injection if the attacker controls the HTTP request method, as demonstrated by inserting CR and LF control characters in the first argument of putrequest(). NOTE: this is similar to CVE-2020-26116. <p>Publish Date: 2020-09-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26137>CVE-2020-26137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26137">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26137</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: 1.25.9</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"urllib3","packageVersion":"1.24.3","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":true,"dependencyTree":"requests:2.20.0;urllib3:1.24.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.25.9","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-26137","vulnerabilityDetails":"urllib3 before 1.25.9 allows CRLF injection if the attacker controls the HTTP request method, as demonstrated by inserting CR and LF control characters in the first argument of putrequest(). NOTE: this is similar to CVE-2020-26116.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26137","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in none any whl cve medium severity vulnerability vulnerable library none any whl http library with thread safe connection pooling file post and more library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy requests none any whl root library x none any whl vulnerable library found in head commit a href found in base branch master vulnerability details before allows crlf injection if the attacker controls the http request method as demonstrated by inserting cr and lf control characters in the first argument of putrequest note this is similar to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree requests isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails before allows crlf injection if the attacker controls the http request method as demonstrated by inserting cr and lf control characters in the first argument of putrequest note this is similar to cve vulnerabilityurl
0
214,354
16,582,971,490
IssuesEvent
2021-05-31 14:18:30
root-project/root
https://api.github.com/repos/root-project/root
closed
Test failures in tutorial-roofit when building without FFTW
bug in:Testing
On my system I get ``` The following tests FAILED: 1061 - tutorial-roofit-rf208_convolution-py (Failed) 1063 - tutorial-roofit-rf210_angularconv-py (Failed) 1064 - tutorial-roofit-rf211_paramconv-py (Failed) 1095 - tutorial-roofit-rf512_wsfactory_oper-py (Failed) ``` The reason seems to be that I'm building without FFTW, and while the macro versions of those tests are correctly vetoed: https://github.com/root-project/root/blob/098dfcf94a3939597488fef9ee128b3d2e1e591f/tutorials/CMakeLists.txt#L191-L198 similar action is missing for the versions written in Python.
1.0
Test failures in tutorial-roofit when building without FFTW - On my system I get ``` The following tests FAILED: 1061 - tutorial-roofit-rf208_convolution-py (Failed) 1063 - tutorial-roofit-rf210_angularconv-py (Failed) 1064 - tutorial-roofit-rf211_paramconv-py (Failed) 1095 - tutorial-roofit-rf512_wsfactory_oper-py (Failed) ``` The reason seems to be that I'm building without FFTW, and while the macro versions of those tests are correctly vetoed: https://github.com/root-project/root/blob/098dfcf94a3939597488fef9ee128b3d2e1e591f/tutorials/CMakeLists.txt#L191-L198 similar action is missing for the versions written in Python.
non_process
test failures in tutorial roofit when building without fftw on my system i get the following tests failed tutorial roofit convolution py failed tutorial roofit angularconv py failed tutorial roofit paramconv py failed tutorial roofit wsfactory oper py failed the reason seems to be that i m building without fftw and while the macro versions of those tests are correctly vetoed similar action is missing for the versions written in python
0
116,104
11,900,999,823
IssuesEvent
2020-03-30 11:41:01
Students-of-the-city-of-Kostroma/Ray-of-hope
https://api.github.com/repos/Students-of-the-city-of-Kostroma/Ray-of-hope
closed
Story-08. Лента постов не соответствует критериям спецификации
Bug Documentation O3 PR4 Specification Sprint 14 Лента постов
Спецификация ["Лента постов"](https://docs.google.com/document/d/1TP_bUWStqtJZPt8JGelRgwGyq9PCXUbF2qKxWsSFwJQ/edit) (версия от 7 декабря 2019 г., 21:58) не соответствует следующим критериям: Критерий | Результат -- | -- Завершённость | + Непротиворечивость | +  Ясность | - Согласованность | + Проверяемость | + Наличие структуры | + Полнота описания | + Логика | + Однозначность | - Единичность | + Комментарий: 13-ый абзац (о фильтрации записей): непонятно, в каких случаях и где присутствует фильтр "Только в моём городе".
1.0
Story-08. Лента постов не соответствует критериям спецификации - Спецификация ["Лента постов"](https://docs.google.com/document/d/1TP_bUWStqtJZPt8JGelRgwGyq9PCXUbF2qKxWsSFwJQ/edit) (версия от 7 декабря 2019 г., 21:58) не соответствует следующим критериям: Критерий | Результат -- | -- Завершённость | + Непротиворечивость | +  Ясность | - Согласованность | + Проверяемость | + Наличие структуры | + Полнота описания | + Логика | + Однозначность | - Единичность | + Комментарий: 13-ый абзац (о фильтрации записей): непонятно, в каких случаях и где присутствует фильтр "Только в моём городе".
non_process
story лента постов не соответствует критериям спецификации спецификация версия от декабря г не соответствует следующим критериям критерий результат завершённость непротиворечивость   ясность согласованность проверяемость наличие структуры полнота описания логика однозначность единичность комментарий ый абзац о фильтрации записей непонятно в каких случаях и где присутствует фильтр только в моём городе
0
70,208
7,179,452,200
IssuesEvent
2018-01-31 19:44:26
vmware/vic
https://api.github.com/repos/vmware/vic
closed
nightly 5-1-Distributed-Switch: Install VCH fails in validation: Post to https://VC_IP/sdk:EOF
component/test kind/nightly-blocker priority/high status/need-info status/needs-estimation team/lifecycle
VCH installation fails in one of the validation steps, with error `Post https://10.162.28.135/sdk: EOF`. `10.162.28.135` is the VC IP according to the log (`--target=https://10.162.28.135`) So this looks like it just lost connection to nimbus all of a sudden. Where exactly in the validation step it fails is hard to know. Sometimes when there's an error during validation, the error is not going to log immediately. They're collected and printed out when validation of everything else finishes. This error falls in this case. All the logs during validation do not show any apparent failure. From `vic-machine.log`: ``` time="2017-12-16T05:48:54-06:00" level=info msg=" \"/ha-datacenter/host/10.160.227.177/10.160.227.177\"" time="2017-12-16T05:48:54-06:00" level=info msg="DRS check SKIPPED - target is standalone host" Dec 16 2017 05:48:54.674-06:00 DEBUG URL: https://harbor.ci.drone.local/v2/ Dec 16 2017 05:48:59.678-06:00 DEBUG URL: http://harbor.ci.drone.local/v2/ time="2017-12-16T05:49:04-06:00" level=warning msg="Unable to confirm insecure registry harbor.ci.drone.local is a valid registry at this time." time="2017-12-16T05:49:04-06:00" level=info msg="Insecure registries = harbor.ci.drone.local" time="2017-12-16T05:49:05-06:00" level=error msg=-------------------- time="2017-12-16T05:49:05-06:00" level=error msg="Post https://10.162.28.135/sdk: EOF" time="2017-12-16T05:49:05-06:00" level=error msg="Create cannot continue: configuration validation failed" time="2017-12-16T05:49:05-06:00" level=error msg=-------------------- time="2017-12-16T05:49:05-06:00" level=error msg="vic-machine-linux create failed: validation of configuration failed\n" ' does not contain 'Installer completed successfully' ``` Log bundle: [5-1-Distributed-Switch.zip](https://github.com/vmware/vic/files/1565340/5-1-Distributed-Switch.zip)
1.0
nightly 5-1-Distributed-Switch: Install VCH fails in validation: Post to https://VC_IP/sdk:EOF - VCH installation fails in one of the validation steps, with error `Post https://10.162.28.135/sdk: EOF`. `10.162.28.135` is the VC IP according to the log (`--target=https://10.162.28.135`) So this looks like it just lost connection to nimbus all of a sudden. Where exactly in the validation step it fails is hard to know. Sometimes when there's an error during validation, the error is not going to log immediately. They're collected and printed out when validation of everything else finishes. This error falls in this case. All the logs during validation do not show any apparent failure. From `vic-machine.log`: ``` time="2017-12-16T05:48:54-06:00" level=info msg=" \"/ha-datacenter/host/10.160.227.177/10.160.227.177\"" time="2017-12-16T05:48:54-06:00" level=info msg="DRS check SKIPPED - target is standalone host" Dec 16 2017 05:48:54.674-06:00 DEBUG URL: https://harbor.ci.drone.local/v2/ Dec 16 2017 05:48:59.678-06:00 DEBUG URL: http://harbor.ci.drone.local/v2/ time="2017-12-16T05:49:04-06:00" level=warning msg="Unable to confirm insecure registry harbor.ci.drone.local is a valid registry at this time." time="2017-12-16T05:49:04-06:00" level=info msg="Insecure registries = harbor.ci.drone.local" time="2017-12-16T05:49:05-06:00" level=error msg=-------------------- time="2017-12-16T05:49:05-06:00" level=error msg="Post https://10.162.28.135/sdk: EOF" time="2017-12-16T05:49:05-06:00" level=error msg="Create cannot continue: configuration validation failed" time="2017-12-16T05:49:05-06:00" level=error msg=-------------------- time="2017-12-16T05:49:05-06:00" level=error msg="vic-machine-linux create failed: validation of configuration failed\n" ' does not contain 'Installer completed successfully' ``` Log bundle: [5-1-Distributed-Switch.zip](https://github.com/vmware/vic/files/1565340/5-1-Distributed-Switch.zip)
non_process
nightly distributed switch install vch fails in validation post to vch installation fails in one of the validation steps with error post eof is the vc ip according to the log target so this looks like it just lost connection to nimbus all of a sudden where exactly in the validation step it fails is hard to know sometimes when there s an error during validation the error is not going to log immediately they re collected and printed out when validation of everything else finishes this error falls in this case all the logs during validation do not show any apparent failure from vic machine log time level info msg ha datacenter host time level info msg drs check skipped target is standalone host dec debug url dec debug url time level warning msg unable to confirm insecure registry harbor ci drone local is a valid registry at this time time level info msg insecure registries harbor ci drone local time level error msg time level error msg post eof time level error msg create cannot continue configuration validation failed time level error msg time level error msg vic machine linux create failed validation of configuration failed n does not contain installer completed successfully log bundle
0
9,018
12,125,509,941
IssuesEvent
2020-04-22 15:38:31
usgs-makerspace/makerspace-sandbox
https://api.github.com/repos/usgs-makerspace/makerspace-sandbox
closed
Calendar events - schedule event and set reminder
Makerspace Processes
We are changing to our new meeting schedule and need to get these added to the calendar Completion Criteria - Add the meeting to calendar and add automated reminder if needed - [x] Marty - standup, demo - [x] Aaron - planning - with reminder - [x] Lindsay - crit - with reminder - [x] Megan - retro - [x] Alicia - expo check, process dev, idea blitz, quarter in review - with reminder - [x] David - demo reminder
1.0
Calendar events - schedule event and set reminder - We are changing to our new meeting schedule and need to get these added to the calendar Completion Criteria - Add the meeting to calendar and add automated reminder if needed - [x] Marty - standup, demo - [x] Aaron - planning - with reminder - [x] Lindsay - crit - with reminder - [x] Megan - retro - [x] Alicia - expo check, process dev, idea blitz, quarter in review - with reminder - [x] David - demo reminder
process
calendar events schedule event and set reminder we are changing to our new meeting schedule and need to get these added to the calendar completion criteria add the meeting to calendar and add automated reminder if needed marty standup demo aaron planning with reminder lindsay crit with reminder megan retro alicia expo check process dev idea blitz quarter in review with reminder david demo reminder
1
21,175
28,144,563,142
IssuesEvent
2023-04-02 10:33:57
log2timeline/plaso
https://api.github.com/repos/log2timeline/plaso
closed
Decouple preprocessing from session
preprocessing
Currently preprocessing is bound to a session, but this does not support: * multiple installation given a single storage media image * process storage media image and additional files related to the same host - [x] Store preprocessing separate from session - ~~store remainder of preprocess information in system configuration artifact~~ - ~~move process configuration into a separate object~~ - https://github.com/log2timeline/plaso/issues/1651 - [x] ~~re-introduce use old preprocess functionality~~
1.0
Decouple preprocessing from session - Currently preprocessing is bound to a session, but this does not support: * multiple installation given a single storage media image * process storage media image and additional files related to the same host - [x] Store preprocessing separate from session - ~~store remainder of preprocess information in system configuration artifact~~ - ~~move process configuration into a separate object~~ - https://github.com/log2timeline/plaso/issues/1651 - [x] ~~re-introduce use old preprocess functionality~~
process
decouple preprocessing from session currently preprocessing is bound to a session but this does not support multiple installation given a single storage media image process storage media image and additional files related to the same host store preprocessing separate from session store remainder of preprocess information in system configuration artifact move process configuration into a separate object re introduce use old preprocess functionality
1
20,236
26,840,577,231
IssuesEvent
2023-02-02 23:56:21
python/cpython
https://api.github.com/repos/python/cpython
closed
multiprocessing's default posix start method of `'fork'` is broken: change to `'spawn'`
type-feature 3.12 expert-multiprocessing
BPO | [40379](https://bugs.python.org/issue40379) --- | :--- Nosy | @pitrou, @mgorny, @Julian, @wimglenn, @applio, @itamarst <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-04-24.18:22:23.389> labels = ['3.8', 'type-bug', '3.7', '3.9'] title = "multiprocessing's default start method of fork()-without-exec() is broken" updated_at = <Date 2022-02-11.16:13:53.872> user = 'https://bugs.python.org/itamarst' ``` bugs.python.org fields: ```python activity = <Date 2022-02-11.16:13:53.872> actor = 'mgorny' assignee = 'none' closed = False closed_date = None closer = None components = [] creation = <Date 2020-04-24.18:22:23.389> creator = 'itamarst' dependencies = [] files = [] hgrepos = [] issue_num = 40379 keywords = [] message_count = 11.0 messages = ['367210', '367211', '368173', '380478', '392358', '392501', '392503', '392506', '392507', '392508', '413081'] nosy_count = 8.0 nosy_names = ['pitrou', 'mgorny', 'Julian', 'wim.glenn', 'itamarst', 'davin', 'itamarst2', 'aduncan'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue40379' versions = ['Python 3.5', 'Python 3.6', 'Python 3.7', 'Python 3.8', 'Python 3.9'] ``` </p></details> <!-- gh-linked-prs --> ### Linked PRs * gh-100618 <!-- /gh-linked-prs -->
1.0
multiprocessing's default posix start method of `'fork'` is broken: change to `'spawn'` - BPO | [40379](https://bugs.python.org/issue40379) --- | :--- Nosy | @pitrou, @mgorny, @Julian, @wimglenn, @applio, @itamarst <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-04-24.18:22:23.389> labels = ['3.8', 'type-bug', '3.7', '3.9'] title = "multiprocessing's default start method of fork()-without-exec() is broken" updated_at = <Date 2022-02-11.16:13:53.872> user = 'https://bugs.python.org/itamarst' ``` bugs.python.org fields: ```python activity = <Date 2022-02-11.16:13:53.872> actor = 'mgorny' assignee = 'none' closed = False closed_date = None closer = None components = [] creation = <Date 2020-04-24.18:22:23.389> creator = 'itamarst' dependencies = [] files = [] hgrepos = [] issue_num = 40379 keywords = [] message_count = 11.0 messages = ['367210', '367211', '368173', '380478', '392358', '392501', '392503', '392506', '392507', '392508', '413081'] nosy_count = 8.0 nosy_names = ['pitrou', 'mgorny', 'Julian', 'wim.glenn', 'itamarst', 'davin', 'itamarst2', 'aduncan'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue40379' versions = ['Python 3.5', 'Python 3.6', 'Python 3.7', 'Python 3.8', 'Python 3.9'] ``` </p></details> <!-- gh-linked-prs --> ### Linked PRs * gh-100618 <!-- /gh-linked-prs -->
process
multiprocessing s default posix start method of fork is broken change to spawn bpo nosy pitrou mgorny julian wimglenn applio itamarst note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title multiprocessing s default start method of fork without exec is broken updated at user bugs python org fields python activity actor mgorny assignee none closed false closed date none closer none components creation creator itamarst dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type behavior url versions linked prs gh
1
2,038
4,847,451,697
IssuesEvent
2016-11-10 15:00:13
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
Cannot cancel a process
browser: all bug comp: activiti-processList
When trying to cancel a process no actions are performed and an error occurs Error in console: `:9999/activiti-app/api/enterprise/process-instances/2547:1 DELETE http://localhost:9999/activiti-app/api/enterprise/process-instances/2547 404 (Not Found)`
1.0
Cannot cancel a process - When trying to cancel a process no actions are performed and an error occurs Error in console: `:9999/activiti-app/api/enterprise/process-instances/2547:1 DELETE http://localhost:9999/activiti-app/api/enterprise/process-instances/2547 404 (Not Found)`
process
cannot cancel a process when trying to cancel a process no actions are performed and an error occurs error in console activiti app api enterprise process instances delete not found
1
93,480
15,891,712,573
IssuesEvent
2021-04-10 20:24:19
HactuallyBenji/321_12
https://api.github.com/repos/HactuallyBenji/321_12
opened
Indicate How Collected Information Will Be Used
functional safety security
Rationale: This application may log sensitive user information. The app must explain why this user info is collected, alongside the ability to consent to fields of information collection. Test Case: 1. Download the application on a new device. 2. Open the application. 3. A popup will appear explaining how collected information will be used, and asking for consent for various fields of data collection. 4. Consent to some of the fields. 5. Attempt to use a feature that needs information from a field that has not been given consent. 6. Verify that a popup is displayed explaining how the collected information will be used, and asking again for consent.
True
Indicate How Collected Information Will Be Used - Rationale: This application may log sensitive user information. The app must explain why this user info is collected, alongside the ability to consent to fields of information collection. Test Case: 1. Download the application on a new device. 2. Open the application. 3. A popup will appear explaining how collected information will be used, and asking for consent for various fields of data collection. 4. Consent to some of the fields. 5. Attempt to use a feature that needs information from a field that has not been given consent. 6. Verify that a popup is displayed explaining how the collected information will be used, and asking again for consent.
non_process
indicate how collected information will be used rationale this application may log sensitive user information the app must explain why this user info is collected alongside the ability to consent to fields of information collection test case download the application on a new device open the application a popup will appear explaining how collected information will be used and asking for consent for various fields of data collection consent to some of the fields attempt to use a feature that needs information from a field that has not been given consent verify that a popup is displayed explaining how the collected information will be used and asking again for consent
0
20,585
27,245,895,410
IssuesEvent
2023-02-22 01:57:30
cse442-at-ub/project_s23-team-infinity
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
opened
Create a Hello World page by using React.js and PHP
Processing Task Sprint 1
**Task Test** *Test 1* 1) run 'npm start' command in the directory 2) Verify that a webpage is running on localhost:3000 by default 3) Verify that there is a 'Hello World' message shows on the webpage 4) Verify that there is a field to enter your name and a submit button 5) Type your name in the field 6) Verify that it holds the functionality of typing in the field 7) Hit submit button 8) Verify that your name shows on the screen properly with the 'Hello World' message 9) Type another name and verity that the message shows correctly with the name you type in the second time.
1.0
Create a Hello World page by using React.js and PHP - **Task Test** *Test 1* 1) run 'npm start' command in the directory 2) Verify that a webpage is running on localhost:3000 by default 3) Verify that there is a 'Hello World' message shows on the webpage 4) Verify that there is a field to enter your name and a submit button 5) Type your name in the field 6) Verify that it holds the functionality of typing in the field 7) Hit submit button 8) Verify that your name shows on the screen properly with the 'Hello World' message 9) Type another name and verity that the message shows correctly with the name you type in the second time.
process
create a hello world page by using react js and php task test test run npm start command in the directory verify that a webpage is running on localhost by default verify that there is a hello world message shows on the webpage verify that there is a field to enter your name and a submit button type your name in the field verify that it holds the functionality of typing in the field hit submit button verify that your name shows on the screen properly with the hello world message type another name and verity that the message shows correctly with the name you type in the second time
1