Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
853
| labels
stringlengths 4
898
| body
stringlengths 2
262k
| index
stringclasses 13
values | text_combine
stringlengths 96
262k
| label
stringclasses 2
values | text
stringlengths 96
250k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
55,002
| 13,495,900,165
|
IssuesEvent
|
2020-09-12 01:24:02
|
GoogleCloudPlatform/java-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/java-docs-samples
|
closed
|
com.example.automl.ListModelsTest: testListModels failed
|
api: automl buildcop: issue priority: p1 samples type: bug
|
Note: #3574 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 5b18a34bf1b60e157c36d2d12a0c15cc753f4135
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/8374a1ca-9ee4-4094-aef6-d69e36caee18), [Sponge](http://sponge2/8374a1ca-9ee4-4094-aef6-d69e36caee18)
status: failed
<details><summary>Test output</summary><br><pre>com.google.api.gax.rpc.DeadlineExceededException: io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: deadline exceeded after 4.999735966s. [buffered_nanos=11744996, buffered_nanos=15874224, remote_addr=automl.googleapis.com/74.125.195.95:443]
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:51)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60)
at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97)
at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1050)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1176)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:969)
at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:760)
at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:434)
at io.grpc.internal.ClientCallImpl.access$500(ClientCallImpl.java:66)
at io.grpc.internal.ClientCallImpl$1CloseInContext.runInContext(ClientCallImpl.java:424)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed
at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57)
at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:112)
at com.google.cloud.automl.v1.AutoMlClient.listModels(AutoMlClient.java:1431)
at com.example.automl.ListModels.listModels(ListModels.java:52)
at com.example.automl.ListModelsTest.testListModels(ListModelsTest.java:64)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451)
Caused by: io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: deadline exceeded after 4.999735966s. [buffered_nanos=11744996, buffered_nanos=15874224, remote_addr=automl.googleapis.com/74.125.195.95:443]
at io.grpc.Status.asRuntimeException(Status.java:533)
... 12 more
</pre></details>
|
1.0
|
com.example.automl.ListModelsTest: testListModels failed - Note: #3574 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 5b18a34bf1b60e157c36d2d12a0c15cc753f4135
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/8374a1ca-9ee4-4094-aef6-d69e36caee18), [Sponge](http://sponge2/8374a1ca-9ee4-4094-aef6-d69e36caee18)
status: failed
<details><summary>Test output</summary><br><pre>com.google.api.gax.rpc.DeadlineExceededException: io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: deadline exceeded after 4.999735966s. [buffered_nanos=11744996, buffered_nanos=15874224, remote_addr=automl.googleapis.com/74.125.195.95:443]
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:51)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60)
at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97)
at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1050)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1176)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:969)
at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:760)
at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:434)
at io.grpc.internal.ClientCallImpl.access$500(ClientCallImpl.java:66)
at io.grpc.internal.ClientCallImpl$1CloseInContext.runInContext(ClientCallImpl.java:424)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed
at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57)
at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:112)
at com.google.cloud.automl.v1.AutoMlClient.listModels(AutoMlClient.java:1431)
at com.example.automl.ListModels.listModels(ListModels.java:52)
at com.example.automl.ListModelsTest.testListModels(ListModelsTest.java:64)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451)
Caused by: io.grpc.StatusRuntimeException: DEADLINE_EXCEEDED: deadline exceeded after 4.999735966s. [buffered_nanos=11744996, buffered_nanos=15874224, remote_addr=automl.googleapis.com/74.125.195.95:443]
at io.grpc.Status.asRuntimeException(Status.java:533)
... 12 more
</pre></details>
|
build
|
com example automl listmodelstest testlistmodels failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output com google api gax rpc deadlineexceededexception io grpc statusruntimeexception deadline exceeded deadline exceeded after at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java base java util concurrent executors runnableadapter call executors java at java base java util concurrent futuretask run futuretask java at java base java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java suppressed com google api gax rpc asynctaskexception asynchronous task failed at com google api gax rpc apiexceptions callandtranslateapiexception apiexceptions java at com google api gax rpc unarycallable call unarycallable java at com google cloud automl automlclient listmodels automlclient java at com example automl listmodels listmodels listmodels java at com example automl listmodelstest testlistmodels listmodelstest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by io grpc statusruntimeexception deadline exceeded deadline exceeded after at io grpc status asruntimeexception status java more
| 1
|
28,966
| 5,447,905,018
|
IssuesEvent
|
2017-03-07 14:44:06
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
reopened
|
Crashed pdns while (bulk) creating zones via the api
|
auth defect
|
I was creating 225 zones in pdns using nsedit, which uses the API. That were quite a lot of requests fired at the API.
It seems (now) that during that import, pdns crashed twice. I didn't notice because it restarted quickly..
nsedit would have been doing the following requests:
- PUT on /servers/localhost/zones
- PATCH on /servers/localhost/zones/<zone> for each non-NS/SOA record (4 in this case)
- Rinse and repeat for next zone
I got the following trace:
```
Nov 6 14:05:38 nscache2 pdns[24922]: Got a signal 6, attempting to print trace:
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x652460]
Nov 6 14:05:38 nscache2 pdns[24922]: /lib/x86_64-linux-gnu/libc.so.6(+0x364a0) [0x7f45d1c384a0]
Nov 6 14:05:38 nscache2 pdns[24922]: /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x35) [0x7f45d1c38425]
Nov 6 14:05:38 nscache2 pdns[24922]: /lib/x86_64-linux-gnu/libc.so.6(abort+0x17b) [0x7f45d1c3bb8b]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(SSQLite3::~SSQLite3()+0xab) [0x72458b]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(SSQLite3::~SSQLite3()+0x9) [0x7245e9]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(GSQLBackend::~GSQLBackend()+0x2e) [0x56c36e]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(gSQLite3Backend::~gSQLite3Backend()+0x17) [0x584c67]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(cleanup_backends(UeberBackend*)+0x26) [0x654246]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(UeberBackend::cleanup()+0x78) [0x654878]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(UeberBackend::~UeberBackend()+0x1d) [0x6548bd]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x635423]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::function2<void, HttpRequest*, HttpResponse*>::operator()(HttpRequest*, HttpResponse*) const+0x18) [0x648af8]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x644bcc]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::detail::function::void_function_obj_invoker2<boost::_bi::bind_t<void, void (*)(boost::function<void (HttpRequest*, HttpResponse*)>, HttpRequest*, HttpResponse*), boost::_bi::list3<boost::_bi::value<boost::function<void (HttpRequest*, HttpResponse*)> >, boost::arg<1>, boost::arg<2> > >, void, HttpRequest*, HttpResponse*>::invoke(boost::detail::function::function_buffer&, HttpRequest*, HttpResponse*)+0x67) [0x6474f7]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::function2<void, HttpRequest*, HttpResponse*>::operator()(HttpRequest*, HttpResponse*) const+0x18) [0x648af8]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::detail::function::void_function_obj_invoker2<boost::_bi::bind_t<void, void (*)(boost::function<void (HttpRequest*, HttpResponse*)>, YaHTTP::Request*, YaHTTP::Response*), boost::_bi::list3<boost::_bi::value<boost::function<void (HttpRequest*, HttpResponse*)> >, boost::arg<1>, boost::arg<2> > >, void, YaHTTP::Request*, YaHTTP::Response*>::invoke(boost::detail::function::function_buffer&, YaHTTP::Request*, YaHTTP::Response*)+0x67) [0x6475d7]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(WebServer::handleRequest(HttpRequest)+0x239) [0x645289]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(WebServer::serveConnection(Socket*)+0x1a5) [0x646775]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x646d92]
Nov 6 14:05:38 nscache2 pdns[12820]: Our pdns instance (24922) exited after signal 6
```
I've got no clue where to look, maybe you guys do.
Running: pdns 0.0.20140723.4983.b6b24bc-1
|
1.0
|
Crashed pdns while (bulk) creating zones via the api - I was creating 225 zones in pdns using nsedit, which uses the API. That were quite a lot of requests fired at the API.
It seems (now) that during that import, pdns crashed twice. I didn't notice because it restarted quickly..
nsedit would have been doing the following requests:
- PUT on /servers/localhost/zones
- PATCH on /servers/localhost/zones/<zone> for each non-NS/SOA record (4 in this case)
- Rinse and repeat for next zone
I got the following trace:
```
Nov 6 14:05:38 nscache2 pdns[24922]: Got a signal 6, attempting to print trace:
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x652460]
Nov 6 14:05:38 nscache2 pdns[24922]: /lib/x86_64-linux-gnu/libc.so.6(+0x364a0) [0x7f45d1c384a0]
Nov 6 14:05:38 nscache2 pdns[24922]: /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x35) [0x7f45d1c38425]
Nov 6 14:05:38 nscache2 pdns[24922]: /lib/x86_64-linux-gnu/libc.so.6(abort+0x17b) [0x7f45d1c3bb8b]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(SSQLite3::~SSQLite3()+0xab) [0x72458b]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(SSQLite3::~SSQLite3()+0x9) [0x7245e9]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(GSQLBackend::~GSQLBackend()+0x2e) [0x56c36e]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(gSQLite3Backend::~gSQLite3Backend()+0x17) [0x584c67]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(cleanup_backends(UeberBackend*)+0x26) [0x654246]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(UeberBackend::cleanup()+0x78) [0x654878]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(UeberBackend::~UeberBackend()+0x1d) [0x6548bd]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x635423]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::function2<void, HttpRequest*, HttpResponse*>::operator()(HttpRequest*, HttpResponse*) const+0x18) [0x648af8]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x644bcc]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::detail::function::void_function_obj_invoker2<boost::_bi::bind_t<void, void (*)(boost::function<void (HttpRequest*, HttpResponse*)>, HttpRequest*, HttpResponse*), boost::_bi::list3<boost::_bi::value<boost::function<void (HttpRequest*, HttpResponse*)> >, boost::arg<1>, boost::arg<2> > >, void, HttpRequest*, HttpResponse*>::invoke(boost::detail::function::function_buffer&, HttpRequest*, HttpResponse*)+0x67) [0x6474f7]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::function2<void, HttpRequest*, HttpResponse*>::operator()(HttpRequest*, HttpResponse*) const+0x18) [0x648af8]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(boost::detail::function::void_function_obj_invoker2<boost::_bi::bind_t<void, void (*)(boost::function<void (HttpRequest*, HttpResponse*)>, YaHTTP::Request*, YaHTTP::Response*), boost::_bi::list3<boost::_bi::value<boost::function<void (HttpRequest*, HttpResponse*)> >, boost::arg<1>, boost::arg<2> > >, void, YaHTTP::Request*, YaHTTP::Response*>::invoke(boost::detail::function::function_buffer&, YaHTTP::Request*, YaHTTP::Response*)+0x67) [0x6475d7]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(WebServer::handleRequest(HttpRequest)+0x239) [0x645289]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance(WebServer::serveConnection(Socket*)+0x1a5) [0x646775]
Nov 6 14:05:38 nscache2 pdns[24922]: /usr/sbin/pdns_server-instance() [0x646d92]
Nov 6 14:05:38 nscache2 pdns[12820]: Our pdns instance (24922) exited after signal 6
```
I've got no clue where to look, maybe you guys do.
Running: pdns 0.0.20140723.4983.b6b24bc-1
|
non_build
|
crashed pdns while bulk creating zones via the api i was creating zones in pdns using nsedit which uses the api that were quite a lot of requests fired at the api it seems now that during that import pdns crashed twice i didn t notice because it restarted quickly nsedit would have been doing the following requests put on servers localhost zones patch on servers localhost zones for each non ns soa record in this case rinse and repeat for next zone i got the following trace nov pdns got a signal attempting to print trace nov pdns usr sbin pdns server instance nov pdns lib linux gnu libc so nov pdns lib linux gnu libc so gsignal nov pdns lib linux gnu libc so abort nov pdns usr sbin pdns server instance nov pdns usr sbin pdns server instance nov pdns usr sbin pdns server instance gsqlbackend gsqlbackend nov pdns usr sbin pdns server instance nov pdns usr sbin pdns server instance cleanup backends ueberbackend nov pdns usr sbin pdns server instance ueberbackend cleanup nov pdns usr sbin pdns server instance ueberbackend ueberbackend nov pdns usr sbin pdns server instance nov pdns usr sbin pdns server instance boost operator httprequest httpresponse const nov pdns usr sbin pdns server instance nov pdns usr sbin pdns server instance boost detail function void function obj httprequest httpresponse boost bi boost arg boost arg void httprequest httpresponse invoke boost detail function function buffer httprequest httpresponse nov pdns usr sbin pdns server instance boost operator httprequest httpresponse const nov pdns usr sbin pdns server instance boost detail function void function obj yahttp request yahttp response boost bi boost arg boost arg void yahttp request yahttp response invoke boost detail function function buffer yahttp request yahttp response nov pdns usr sbin pdns server instance webserver handlerequest httprequest nov pdns usr sbin pdns server instance webserver serveconnection socket nov pdns usr sbin pdns server instance nov pdns our pdns instance exited after signal i ve got no clue where to look maybe you guys do running pdns
| 0
|
333,901
| 29,820,077,288
|
IssuesEvent
|
2023-06-17 00:50:23
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
closed
|
Fix tensor.test_torch_instance_reshape
|
PyTorch Frontend Sub Task Failing Test
|
| | |
|---|---|
|numpy|<a href="null"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix tensor.test_torch_instance_reshape - | | |
|---|---|
|numpy|<a href="null"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5262201861/jobs/9511159860"><img src=https://img.shields.io/badge/-success-success></a>
|
non_build
|
fix tensor test torch instance reshape numpy img src torch a href src tensorflow a href src jax a href src paddle a href src
| 0
|
87,084
| 25,026,757,105
|
IssuesEvent
|
2022-11-04 08:46:57
|
mautic/mautic
|
https://api.github.com/repos/mautic/mautic
|
closed
|
Email editor scrolls/jumps down after every `enter`
|
bug builder-grapesjs
|
### Mautic Version
4.0.0
### PHP version
7.4.26
### What browsers are you seeing the problem on?
_No response_
### What happened?
While editing long emails (MJML template), the editor jumps/scrolls down every time `enter` gets typed.
This isn't a problem if you don't have a vertical scroll bar or (strangely) if you are at the top of the page/email.
If you scroll a bit to edit text in the middle the editor starts to jump on every `enter`.
### How can we reproduce this issue?
1. Open the email editor of a long email
2. Scroll to the middle
3. Edit a text and hit `enter`
### Relevant log output
```shell
There is none
```
### Code of Conduct
- [X] I confirm that I have read and agree to follow this project's Code of Conduct
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://app.bountysource.com/issues/108497649-email-editor-scrolls-jumps-down-after-every-enter?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://app.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
Email editor scrolls/jumps down after every `enter` - ### Mautic Version
4.0.0
### PHP version
7.4.26
### What browsers are you seeing the problem on?
_No response_
### What happened?
While editing long emails (MJML template), the editor jumps/scrolls down every time `enter` gets typed.
This isn't a problem if you don't have a vertical scroll bar or (strangely) if you are at the top of the page/email.
If you scroll a bit to edit text in the middle the editor starts to jump on every `enter`.
### How can we reproduce this issue?
1. Open the email editor of a long email
2. Scroll to the middle
3. Edit a text and hit `enter`
### Relevant log output
```shell
There is none
```
### Code of Conduct
- [X] I confirm that I have read and agree to follow this project's Code of Conduct
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://app.bountysource.com/issues/108497649-email-editor-scrolls-jumps-down-after-every-enter?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://app.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
build
|
email editor scrolls jumps down after every enter mautic version php version what browsers are you seeing the problem on no response what happened while editing long emails mjml template the editor jumps scrolls down every time enter gets typed this isn t a problem if you don t have a vertical scroll bar or strangely if you are at the top of the page email if you scroll a bit to edit text in the middle the editor starts to jump on every enter how can we reproduce this issue open the email editor of a long email scroll to the middle edit a text and hit enter relevant log output shell there is none code of conduct i confirm that i have read and agree to follow this project s code of conduct want to back this issue we accept bounties via
| 1
|
247,601
| 20,987,263,055
|
IssuesEvent
|
2022-03-29 05:27:21
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: clock/jump/large_backward_disabled failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker
|
roachtest.clock/jump/large_backward_disabled [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=artifacts#/clock/jump/large_backward_disabled) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /artifacts/clock/jump/large_backward_disabled/run_1
cluster.go:1868,clock_jump_crash.go:43,clock_jump_crash.go:143,test_runner.go:875: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1826
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1867
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:43
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerClockJumpTests.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:143
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:875
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clock/jump/large_backward_disabled.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: clock/jump/large_backward_disabled failed - roachtest.clock/jump/large_backward_disabled [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=artifacts#/clock/jump/large_backward_disabled) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /artifacts/clock/jump/large_backward_disabled/run_1
cluster.go:1868,clock_jump_crash.go:43,clock_jump_crash.go:143,test_runner.go:875: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1826
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1867
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:43
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerClockJumpTests.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:143
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:875
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clock/jump/large_backward_disabled.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
non_build
|
roachtest clock jump large backward disabled failed roachtest clock jump large backward disabled with on master the test failed on branch master cloud gce test artifacts and logs in artifacts clock jump large backward disabled run cluster go clock jump crash go clock jump crash go test runner go one or more parallel execution failure attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster parallele github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster parallel github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster start github com cockroachdb cockroach pkg roachprod install cockroach go github com cockroachdb cockroach pkg roachprod start github com cockroachdb cockroach pkg roachprod roachprod go main clusterimpl starte main pkg cmd roachtest cluster go main clusterimpl start main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests runclockjump github com cockroachdb cockroach pkg cmd roachtest tests clock jump crash go github com cockroachdb cockroach pkg cmd roachtest tests registerclockjumptests github com cockroachdb cockroach pkg cmd roachtest tests clock jump crash go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps one or more parallel execution failure error types withstack withstack errutil leaferror help see see cc cockroachdb kv triage
| 0
|
11,137
| 2,639,313,824
|
IssuesEvent
|
2015-03-11 00:19:16
|
kronometrix/recording
|
https://api.github.com/repos/kronometrix/recording
|
closed
|
pkg removal does not stop data recorders
|
defect-high
|
After purging the deb package we can still see data recorders running:
```
# dpkg -P kronometrix
root 3353 1 0 Nov19 ? 00:03:51 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/sysrec 60
root 3370 1 0 Nov19 ? 00:00:26 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/cpurec 60
root 3387 1 0 Nov19 ? 00:02:22 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/diskrec 60
root 3404 1 0 Nov19 ? 00:00:46 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/nicrec 60
root 3421 1 0 Nov19 ? 00:00:00 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/hdwrec 86400
root 4440 1 0 Nov20 ? 00:00:35 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/wsrec -p 60
```
|
1.0
|
pkg removal does not stop data recorders - After purging the deb package we can still see data recorders running:
```
# dpkg -P kronometrix
root 3353 1 0 Nov19 ? 00:03:51 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/sysrec 60
root 3370 1 0 Nov19 ? 00:00:26 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/cpurec 60
root 3387 1 0 Nov19 ? 00:02:22 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/diskrec 60
root 3404 1 0 Nov19 ? 00:00:46 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/nicrec 60
root 3421 1 0 Nov19 ? 00:00:00 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/hdwrec 86400
root 4440 1 0 Nov20 ? 00:00:35 /opt/kronometrix/perl/bin/perl /opt/kronometrix/bin/wsrec -p 60
```
|
non_build
|
pkg removal does not stop data recorders after purging the deb package we can still see data recorders running dpkg p kronometrix root opt kronometrix perl bin perl opt kronometrix bin sysrec root opt kronometrix perl bin perl opt kronometrix bin cpurec root opt kronometrix perl bin perl opt kronometrix bin diskrec root opt kronometrix perl bin perl opt kronometrix bin nicrec root opt kronometrix perl bin perl opt kronometrix bin hdwrec root opt kronometrix perl bin perl opt kronometrix bin wsrec p
| 0
|
59,728
| 17,023,222,242
|
IssuesEvent
|
2021-07-03 00:55:48
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Relations enhancements
|
Component: potlatch (flash editor) Priority: major Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 11.57am, Wednesday, 19th March 2008]**
- Splitting a way with relation should apply the relation to both parts
- Extending a way with relation doesn't repaint properly
- Repeat!
|
1.0
|
Relations enhancements - **[Submitted to the original trac issue database at 11.57am, Wednesday, 19th March 2008]**
- Splitting a way with relation should apply the relation to both parts
- Extending a way with relation doesn't repaint properly
- Repeat!
|
non_build
|
relations enhancements splitting a way with relation should apply the relation to both parts extending a way with relation doesn t repaint properly repeat
| 0
|
25,201
| 5,143,347,333
|
IssuesEvent
|
2017-01-12 15:47:21
|
socketio/socket.io
|
https://api.github.com/repos/socketio/socket.io
|
closed
|
Usage XHR vs JSONP docs?
|
Closed due to inactivity Documentation
|
I can't find information on when JSONP is used and when XHR is used.
I checked the readme's from:
- this repository
- engine.io-protocol
- engine.io
- engine.io-client
I think I understand one of these is used when a client is in "polling" mode, but there is little information on why one of these would be selected.
I understand I can force JSONP using the `forceJSONP` option and I could disable it by setting `jsonp` to false.
|
1.0
|
Usage XHR vs JSONP docs? - I can't find information on when JSONP is used and when XHR is used.
I checked the readme's from:
- this repository
- engine.io-protocol
- engine.io
- engine.io-client
I think I understand one of these is used when a client is in "polling" mode, but there is little information on why one of these would be selected.
I understand I can force JSONP using the `forceJSONP` option and I could disable it by setting `jsonp` to false.
|
non_build
|
usage xhr vs jsonp docs i can t find information on when jsonp is used and when xhr is used i checked the readme s from this repository engine io protocol engine io engine io client i think i understand one of these is used when a client is in polling mode but there is little information on why one of these would be selected i understand i can force jsonp using the forcejsonp option and i could disable it by setting jsonp to false
| 0
|
25,356
| 7,693,781,968
|
IssuesEvent
|
2018-05-18 06:00:42
|
h2oai/h2o4gpu
|
https://api.github.com/repos/h2oai/h2o4gpu
|
closed
|
CUDA 9.1 support
|
build
|
CUDA 9.1 is not backwards compatible with 9.0. Currently it is failing to build/run with several errors. Need to look into why exactly and how to fix them.
We also need to figure out if/how we want to support both 9.0 and 9.1 at the same time.
|
1.0
|
CUDA 9.1 support - CUDA 9.1 is not backwards compatible with 9.0. Currently it is failing to build/run with several errors. Need to look into why exactly and how to fix them.
We also need to figure out if/how we want to support both 9.0 and 9.1 at the same time.
|
build
|
cuda support cuda is not backwards compatible with currently it is failing to build run with several errors need to look into why exactly and how to fix them we also need to figure out if how we want to support both and at the same time
| 1
|
367,516
| 25,745,658,842
|
IssuesEvent
|
2022-12-08 09:40:02
|
knowease-inc/knowease-inc.github.io
|
https://api.github.com/repos/knowease-inc/knowease-inc.github.io
|
opened
|
쉬운지식 -> (주)게타 변경
|
Domain:Business ETR:1W- Priority:High Task:Documentation
|
## 이런 목표를 달성해야 합니다
> 이 이슈로 무슨 목표를 달성하고자 하며 어떤 상태가 되어야 하는지 간결히 적어주세요.
- (주)게타로 명칭 변경
## 현재 이런 상태입니다
> 이 이슈를 생성한 현시점의 문제 혹은 향후 문제 발생 가능성에 대하여 간결히 적어주세요.
- 필요에 의해 주식회사 게타를 신규설립했습니다
## 이 이슈는 이 분이 풀 수 있을 것 같습니다
> 담당할 Assignee를 @로 **1명만** 멘션해주세요.
- @T-Mook
## 아래의 세부적인 문제를 풀어야 할 것 같습니다
> 이 이슈를 해결하기 위한 세부 항목(이슈 클로징 조건)을 체크리스트로 적어주세요.
- [ ] 변경
- [ ] 확인
## 이 이슈를 해결하기 위해 이런 내용을 참고할 수 있을 것 같습니다
> 문제 해결에 도움이 될 수 있을 것 같은 관련 이슈 번호, 문서, Wiki, 스크린샷, 개인적인 의견 등을 최대한 적어주세요.
> 이 이슈가 다른 이슈와 관련되어 있는 경우는 **반드시 이슈 번호를 적어주세요**
- 관련이슈: https://github.com/knowease-inc/Business-Strategies/issues/160
## 이 이슈 해결을 위해 이정도 시간이 예상됩니다
> 예상소요시간을 한가지만 선택해주세요.
> (1W+ 가 아닌 경우 레이블을 변경해주세요.)
- 예상소요시간: **1W-**
## 관련된 세부 정보입니다.
> Reporter는 **1명만**, Domain, Priority, Task를 **각각 한가지만** 선택해주세요.
> (UX, Medium, Enhancement 가 아닌 경우 레이블을 변경해주세요.)
- Reporter: @T-Mook
- Domain : **Business**
- Priority: **High**
- Task : **Documentation**
## 이 이슈를 해결함에 따라 이정도 재무적 영향이 예상됩니다.
> 이 이슈를 해결함에 따라 전사적으로 유의미한 수익/비용 변동이 예상될 경우, 해당 수치를 입력해주세요.
- 예상수익: 0 원/월
- 예상비용: 0 원/월
|
1.0
|
쉬운지식 -> (주)게타 변경 - ## 이런 목표를 달성해야 합니다
> 이 이슈로 무슨 목표를 달성하고자 하며 어떤 상태가 되어야 하는지 간결히 적어주세요.
- (주)게타로 명칭 변경
## 현재 이런 상태입니다
> 이 이슈를 생성한 현시점의 문제 혹은 향후 문제 발생 가능성에 대하여 간결히 적어주세요.
- 필요에 의해 주식회사 게타를 신규설립했습니다
## 이 이슈는 이 분이 풀 수 있을 것 같습니다
> 담당할 Assignee를 @로 **1명만** 멘션해주세요.
- @T-Mook
## 아래의 세부적인 문제를 풀어야 할 것 같습니다
> 이 이슈를 해결하기 위한 세부 항목(이슈 클로징 조건)을 체크리스트로 적어주세요.
- [ ] 변경
- [ ] 확인
## 이 이슈를 해결하기 위해 이런 내용을 참고할 수 있을 것 같습니다
> 문제 해결에 도움이 될 수 있을 것 같은 관련 이슈 번호, 문서, Wiki, 스크린샷, 개인적인 의견 등을 최대한 적어주세요.
> 이 이슈가 다른 이슈와 관련되어 있는 경우는 **반드시 이슈 번호를 적어주세요**
- 관련이슈: https://github.com/knowease-inc/Business-Strategies/issues/160
## 이 이슈 해결을 위해 이정도 시간이 예상됩니다
> 예상소요시간을 한가지만 선택해주세요.
> (1W+ 가 아닌 경우 레이블을 변경해주세요.)
- 예상소요시간: **1W-**
## 관련된 세부 정보입니다.
> Reporter는 **1명만**, Domain, Priority, Task를 **각각 한가지만** 선택해주세요.
> (UX, Medium, Enhancement 가 아닌 경우 레이블을 변경해주세요.)
- Reporter: @T-Mook
- Domain : **Business**
- Priority: **High**
- Task : **Documentation**
## 이 이슈를 해결함에 따라 이정도 재무적 영향이 예상됩니다.
> 이 이슈를 해결함에 따라 전사적으로 유의미한 수익/비용 변동이 예상될 경우, 해당 수치를 입력해주세요.
- 예상수익: 0 원/월
- 예상비용: 0 원/월
|
non_build
|
쉬운지식 주 게타 변경 이런 목표를 달성해야 합니다 이 이슈로 무슨 목표를 달성하고자 하며 어떤 상태가 되어야 하는지 간결히 적어주세요 주 게타로 명칭 변경 현재 이런 상태입니다 이 이슈를 생성한 현시점의 문제 혹은 향후 문제 발생 가능성에 대하여 간결히 적어주세요 필요에 의해 주식회사 게타를 신규설립했습니다 이 이슈는 이 분이 풀 수 있을 것 같습니다 담당할 assignee를 로 멘션해주세요 t mook 아래의 세부적인 문제를 풀어야 할 것 같습니다 이 이슈를 해결하기 위한 세부 항목 이슈 클로징 조건 을 체크리스트로 적어주세요 변경 확인 이 이슈를 해결하기 위해 이런 내용을 참고할 수 있을 것 같습니다 문제 해결에 도움이 될 수 있을 것 같은 관련 이슈 번호 문서 wiki 스크린샷 개인적인 의견 등을 최대한 적어주세요 이 이슈가 다른 이슈와 관련되어 있는 경우는 반드시 이슈 번호를 적어주세요 관련이슈 이 이슈 해결을 위해 이정도 시간이 예상됩니다 예상소요시간을 한가지만 선택해주세요 가 아닌 경우 레이블을 변경해주세요 예상소요시간 관련된 세부 정보입니다 reporter는 domain priority task를 각각 한가지만 선택해주세요 ux medium enhancement 가 아닌 경우 레이블을 변경해주세요 reporter t mook domain business priority high task documentation 이 이슈를 해결함에 따라 이정도 재무적 영향이 예상됩니다 이 이슈를 해결함에 따라 전사적으로 유의미한 수익 비용 변동이 예상될 경우 해당 수치를 입력해주세요 예상수익 원 월 예상비용 원 월
| 0
|
338,236
| 10,226,337,466
|
IssuesEvent
|
2019-08-16 17:32:17
|
googleapis/google-api-ruby-client
|
https://api.github.com/repos/googleapis/google-api-ruby-client
|
opened
|
Veneers backed by google-api-client may duplicate "gl-ruby" in the x-goog-api-client header
|
priority: p2 type: bug
|
Affected libraries are google-cloud-bigquery, google-cloud-dns, google-cloud-resource_manager, and google-cloud-storage.
This would happen because the veneer sets both gl-ruby/ and gccl/, and then google-api-client appends both gl-ruby/ and gdcl/, resulting in gl-ruby/ appearing twice. We probably should respond by examining any existing x-goog-api-client, and not adding gl-ruby/ if it is already present.
|
1.0
|
Veneers backed by google-api-client may duplicate "gl-ruby" in the x-goog-api-client header - Affected libraries are google-cloud-bigquery, google-cloud-dns, google-cloud-resource_manager, and google-cloud-storage.
This would happen because the veneer sets both gl-ruby/ and gccl/, and then google-api-client appends both gl-ruby/ and gdcl/, resulting in gl-ruby/ appearing twice. We probably should respond by examining any existing x-goog-api-client, and not adding gl-ruby/ if it is already present.
|
non_build
|
veneers backed by google api client may duplicate gl ruby in the x goog api client header affected libraries are google cloud bigquery google cloud dns google cloud resource manager and google cloud storage this would happen because the veneer sets both gl ruby and gccl and then google api client appends both gl ruby and gdcl resulting in gl ruby appearing twice we probably should respond by examining any existing x goog api client and not adding gl ruby if it is already present
| 0
|
90,859
| 18,267,651,950
|
IssuesEvent
|
2021-10-04 10:18:40
|
google/android-fhir
|
https://api.github.com/repos/google/android-fhir
|
opened
|
Refactor DataCaptureConfig to a more standard android way like Configuration Provider.
|
code health
|
**Describe the Issue**
Based on the [#754 review comment](https://github.com/google/android-fhir/pull/754#pullrequestreview-767105098), make the DataCaptureConfig change to a provider like work manager.
**Would you like to work on the issue?**
|
1.0
|
Refactor DataCaptureConfig to a more standard android way like Configuration Provider. - **Describe the Issue**
Based on the [#754 review comment](https://github.com/google/android-fhir/pull/754#pullrequestreview-767105098), make the DataCaptureConfig change to a provider like work manager.
**Would you like to work on the issue?**
|
non_build
|
refactor datacaptureconfig to a more standard android way like configuration provider describe the issue based on the make the datacaptureconfig change to a provider like work manager would you like to work on the issue
| 0
|
387,211
| 26,716,174,232
|
IssuesEvent
|
2023-01-28 14:45:34
|
endojs/endo
|
https://api.github.com/repos/endojs/endo
|
opened
|
Doc links into agoric.com are 404'ing
|
documentation
|
## Description
Link to https://agoric.com/documentation/js-programming/far.html from [packages/far/README.md](../blob/master/packages/far/README.md) is 404'ing.
|
1.0
|
Doc links into agoric.com are 404'ing - ## Description
Link to https://agoric.com/documentation/js-programming/far.html from [packages/far/README.md](../blob/master/packages/far/README.md) is 404'ing.
|
non_build
|
doc links into agoric com are ing description link to from blob master packages far readme md is ing
| 0
|
75,416
| 20,798,110,142
|
IssuesEvent
|
2022-03-17 11:17:20
|
opentap/opentap
|
https://api.github.com/repos/opentap/opentap
|
opened
|
Visual studio: Misbehaving failing image install
|
bug BUILD
|
When an image install fails, it keeps retrying in an infinite loop. This is because the image install task checks if all the required package.xml files exist on the disk as a condition for running, but obviously the package.xml file does not exist if the image failed to install.
This can be pretty annoying for anyone using Visual Studio, as the errors will keep popping up until the image successfully resolves.
|
1.0
|
Visual studio: Misbehaving failing image install - When an image install fails, it keeps retrying in an infinite loop. This is because the image install task checks if all the required package.xml files exist on the disk as a condition for running, but obviously the package.xml file does not exist if the image failed to install.
This can be pretty annoying for anyone using Visual Studio, as the errors will keep popping up until the image successfully resolves.
|
build
|
visual studio misbehaving failing image install when an image install fails it keeps retrying in an infinite loop this is because the image install task checks if all the required package xml files exist on the disk as a condition for running but obviously the package xml file does not exist if the image failed to install this can be pretty annoying for anyone using visual studio as the errors will keep popping up until the image successfully resolves
| 1
|
108,215
| 23,579,389,881
|
IssuesEvent
|
2022-08-23 06:06:56
|
sast-automation-dev/easybuggy4kt-11
|
https://api.github.com/repos/sast-automation-dev/easybuggy4kt-11
|
opened
|
Code Security Report: 19 high severity findings, 38 total findings
|
code security findings
|
# Code Security Report
**Latest Scan:** 2022-08-23 06:06am
**Total Findings:** 38
**Tested Project Files:** 99
**Detected Programming Languages:** 2
<!-- SAST-MANUAL-SCAN-START -->
- [ ] Check this box to manually trigger a scan
<!-- SAST-MANUAL-SCAN-END -->
## Language: Java
> No vulnerability findings detected.
## Language: Kotlin
| Severity | CWE | Vulnerability Type | Count |
|-|-|-|-|
|<img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High|[CWE-79](https://cwe.mitre.org/data/definitions/79.html)|Cross-Site Scripting|19|
|<img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium|[CWE-338](https://cwe.mitre.org/data/definitions/338.html)|Weak Pseudo-Random|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium|[CWE-244](https://cwe.mitre.org/data/definitions/244.html)|Heap Inspection|7|
|<img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium|[CWE-501](https://cwe.mitre.org/data/definitions/501.html)|Trust Boundary Violation|4|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-434](https://cwe.mitre.org/data/definitions/434.html)|File Upload|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-601](https://cwe.mitre.org/data/definitions/601.html)|Unvalidated/Open Redirect|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-117](https://cwe.mitre.org/data/definitions/117.html)|Log Forging|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-113](https://cwe.mitre.org/data/definitions/113.html)|HTTP Header Injection|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-20](https://cwe.mitre.org/data/definitions/20.html)|Mail Relay|3|
### Details
> The below list presents the 19 high vulnerability findings that need your attention. To view information on these findings, navigate to the [Mend SAST Application](https://dev.whitesourcesoftware.com/sast/#/scans/788b78ae-893a-45a3-9f00-4411b746ddeb/details).
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20>Cross-Site Scripting (CWE-79) : 19</summary>
#### Findings
<details>
<summary>vulnerabilities/XEEandXXEController.kt:47</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XEEandXXEController.kt#L42-L47
</details>
<details>
<summary>vulnerabilities/OGNLExpressionInjectionController.kt:42</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/OGNLExpressionInjectionController.kt#L37-L42
</details>
<details>
<summary>vulnerabilities/XSSController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XSSController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XSSController.kt#L21
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XSSController.kt#L22
</details>
</details>
<details>
<summary>vulnerabilities/XEEandXXEController.kt:53</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XEEandXXEController.kt#L48-L53
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:44</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L39-L44
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L41
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L44
</details>
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:44</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L39-L44
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L32
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L44
</details>
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:44</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L39-L44
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L31
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L44
</details>
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L22
</details>
</details>
<details>
<summary>performance/StringPlusOperationController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L20
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L21
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L22
</details>
</details>
<details>
<summary>vulnerabilities/ClickJackingController.kt:56</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/ClickJackingController.kt#L51-L56
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/ClickJackingController.kt#L48
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/ClickJackingController.kt#L56
</details>
</details>
<details>
<summary>vulnerabilities/CSRFController.kt:59</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/CSRFController.kt#L54-L59
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/CSRFController.kt#L40
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/CSRFController.kt#L59
</details>
</details>
<details>
<summary>troubles/TruncationErrorController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L19
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L22
</details>
</details>
<details>
<summary>troubles/TruncationErrorController.kt:21</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L16-L21
</details>
<details>
<summary>troubles/LossOfTrailingDigitsController.kt:21</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L16-L21
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L18
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L21
</details>
</details>
<details>
<summary>troubles/LossOfTrailingDigitsController.kt:20</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L15-L20
</details>
<details>
<summary>troubles/RoundOffErrorController.kt:21</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L16-L21
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L18
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L21
</details>
</details>
<details>
<summary>troubles/RoundOffErrorController.kt:20</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L15-L20
</details>
<details>
<summary>troubles/IntegerOverflowController.kt:35</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/IntegerOverflowController.kt#L30-L35
</details>
<details>
<summary>vulnerabilities/XEEandXXEController.kt:51</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XEEandXXEController.kt#L46-L51
</details>
</details>
|
1.0
|
Code Security Report: 19 high severity findings, 38 total findings - # Code Security Report
**Latest Scan:** 2022-08-23 06:06am
**Total Findings:** 38
**Tested Project Files:** 99
**Detected Programming Languages:** 2
<!-- SAST-MANUAL-SCAN-START -->
- [ ] Check this box to manually trigger a scan
<!-- SAST-MANUAL-SCAN-END -->
## Language: Java
> No vulnerability findings detected.
## Language: Kotlin
| Severity | CWE | Vulnerability Type | Count |
|-|-|-|-|
|<img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High|[CWE-79](https://cwe.mitre.org/data/definitions/79.html)|Cross-Site Scripting|19|
|<img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium|[CWE-338](https://cwe.mitre.org/data/definitions/338.html)|Weak Pseudo-Random|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium|[CWE-244](https://cwe.mitre.org/data/definitions/244.html)|Heap Inspection|7|
|<img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium|[CWE-501](https://cwe.mitre.org/data/definitions/501.html)|Trust Boundary Violation|4|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-434](https://cwe.mitre.org/data/definitions/434.html)|File Upload|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-601](https://cwe.mitre.org/data/definitions/601.html)|Unvalidated/Open Redirect|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-117](https://cwe.mitre.org/data/definitions/117.html)|Log Forging|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-113](https://cwe.mitre.org/data/definitions/113.html)|HTTP Header Injection|1|
|<img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low|[CWE-20](https://cwe.mitre.org/data/definitions/20.html)|Mail Relay|3|
### Details
> The below list presents the 19 high vulnerability findings that need your attention. To view information on these findings, navigate to the [Mend SAST Application](https://dev.whitesourcesoftware.com/sast/#/scans/788b78ae-893a-45a3-9f00-4411b746ddeb/details).
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20>Cross-Site Scripting (CWE-79) : 19</summary>
#### Findings
<details>
<summary>vulnerabilities/XEEandXXEController.kt:47</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XEEandXXEController.kt#L42-L47
</details>
<details>
<summary>vulnerabilities/OGNLExpressionInjectionController.kt:42</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/OGNLExpressionInjectionController.kt#L37-L42
</details>
<details>
<summary>vulnerabilities/XSSController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XSSController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XSSController.kt#L21
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XSSController.kt#L22
</details>
</details>
<details>
<summary>vulnerabilities/XEEandXXEController.kt:53</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XEEandXXEController.kt#L48-L53
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:44</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L39-L44
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L41
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L44
</details>
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:44</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L39-L44
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L32
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L44
</details>
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:44</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L39-L44
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L31
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L44
</details>
</details>
<details>
<summary>performance/CreatingUnnecessaryObjectsController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L17
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/CreatingUnnecessaryObjectsController.kt#L22
</details>
</details>
<details>
<summary>performance/StringPlusOperationController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L20
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L21
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/performance/StringPlusOperationController.kt#L22
</details>
</details>
<details>
<summary>vulnerabilities/ClickJackingController.kt:56</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/ClickJackingController.kt#L51-L56
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/ClickJackingController.kt#L48
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/ClickJackingController.kt#L56
</details>
</details>
<details>
<summary>vulnerabilities/CSRFController.kt:59</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/CSRFController.kt#L54-L59
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/CSRFController.kt#L40
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/CSRFController.kt#L59
</details>
</details>
<details>
<summary>troubles/TruncationErrorController.kt:22</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L17-L22
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L19
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L22
</details>
</details>
<details>
<summary>troubles/TruncationErrorController.kt:21</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/TruncationErrorController.kt#L16-L21
</details>
<details>
<summary>troubles/LossOfTrailingDigitsController.kt:21</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L16-L21
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L18
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L21
</details>
</details>
<details>
<summary>troubles/LossOfTrailingDigitsController.kt:20</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/LossOfTrailingDigitsController.kt#L15-L20
</details>
<details>
<summary>troubles/RoundOffErrorController.kt:21</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L16-L21
<details>
<summary> Trace </summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L18
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L21
</details>
</details>
<details>
<summary>troubles/RoundOffErrorController.kt:20</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/RoundOffErrorController.kt#L15-L20
</details>
<details>
<summary>troubles/IntegerOverflowController.kt:35</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/troubles/IntegerOverflowController.kt#L30-L35
</details>
<details>
<summary>vulnerabilities/XEEandXXEController.kt:51</summary>
https://github.com/sast-automation-dev/easybuggy4kt-11/blob/a4b6e96e2a731a952676dea649d3b01f116ee86d/easybuggy4kt-11/src/main/kotlin/org/t246osslab/easybuggy4kt/vulnerabilities/XEEandXXEController.kt#L46-L51
</details>
</details>
|
non_build
|
code security report high severity findings total findings code security report latest scan total findings tested project files detected programming languages check this box to manually trigger a scan language java no vulnerability findings detected language kotlin severity cwe vulnerability type count high scripting medium pseudo random medium inspection medium boundary violation low upload low redirect low forging low header injection low relay details the below list presents the high vulnerability findings that need your attention to view information on these findings navigate to the cross site scripting cwe findings vulnerabilities xeeandxxecontroller kt vulnerabilities ognlexpressioninjectioncontroller kt vulnerabilities xsscontroller kt trace vulnerabilities xeeandxxecontroller kt performance creatingunnecessaryobjectscontroller kt trace performance creatingunnecessaryobjectscontroller kt trace performance creatingunnecessaryobjectscontroller kt trace performance creatingunnecessaryobjectscontroller kt trace performance stringplusoperationcontroller kt trace vulnerabilities clickjackingcontroller kt trace vulnerabilities csrfcontroller kt trace troubles truncationerrorcontroller kt trace troubles truncationerrorcontroller kt troubles lossoftrailingdigitscontroller kt trace troubles lossoftrailingdigitscontroller kt troubles roundofferrorcontroller kt trace troubles roundofferrorcontroller kt troubles integeroverflowcontroller kt vulnerabilities xeeandxxecontroller kt
| 0
|
190,717
| 6,821,821,532
|
IssuesEvent
|
2017-11-07 17:57:05
|
CS2103AUG2017-W09-B3/main
|
https://api.github.com/repos/CS2103AUG2017-W09-B3/main
|
closed
|
As a careless user, I want to get a list of contact that I removed (like a Recycle bin)
|
priority.medium status.completed type.enhancement
|
...so that recover the accidentally deleted ones.
- Command class implementation
- Update Parser
- JUnit test
|
1.0
|
As a careless user, I want to get a list of contact that I removed (like a Recycle bin) - ...so that recover the accidentally deleted ones.
- Command class implementation
- Update Parser
- JUnit test
|
non_build
|
as a careless user i want to get a list of contact that i removed like a recycle bin so that recover the accidentally deleted ones command class implementation update parser junit test
| 0
|
14,392
| 3,398,034,355
|
IssuesEvent
|
2015-12-02 00:46:11
|
exercism/todo
|
https://api.github.com/repos/exercism/todo
|
closed
|
Extract shared inputs/outputs for exercise Atbash Cipher
|
shared-test-data
|
Goal: https://github.com/exercism/todo/issues/13
--------------
**Skillset:** This is mostly a matter of reading code (test suites) in several languages and creating some JSON.
---------------
Create a file `atbash-cipher.json` containing the inputs and expected outputs for Atbash Cipher.
Atbash Cipher has been implemented in the following languages:
- https://github.com/exercism/xclojure/tree/master/atbash-cipher
- https://github.com/exercism/xcoffeescript/tree/master/atbash-cipher
- https://github.com/exercism/xcsharp/tree/master/atbash-cipher
- https://github.com/exercism/xelixir/tree/master/atbash-cipher
- https://github.com/exercism/xerlang/tree/master/atbash-cipher
- https://github.com/exercism/xfsharp/tree/master/atbash-cipher
- https://github.com/exercism/xgo/tree/master/atbash-cipher
- https://github.com/exercism/xhaskell/tree/master/atbash-cipher
- https://github.com/exercism/xjava/tree/master/atbash-cipher
- https://github.com/exercism/xjavascript/tree/master/atbash-cipher
- https://github.com/exercism/xlfe/tree/master/atbash-cipher
- https://github.com/exercism/xlisp/tree/master/atbash-cipher
- https://github.com/exercism/xperl5/tree/master/atbash-cipher
- https://github.com/exercism/xpython/tree/master/atbash-cipher
- https://github.com/exercism/xruby/tree/master/atbash-cipher
- https://github.com/exercism/xscala/tree/master/atbash-cipher
- https://github.com/exercism/xswift/tree/master/atbash-cipher
See the following files for some examples:
- https://github.com/exercism/x-common/blob/master/bob.json
- https://github.com/exercism/x-common/blob/master/clock.json
- https://github.com/exercism/x-common/blob/master/custom-set.json
- https://github.com/exercism/x-common/blob/master/gigasecond.json
- https://github.com/exercism/x-common/blob/master/hamming.json
- https://github.com/exercism/x-common/blob/master/leap.json
|
1.0
|
Extract shared inputs/outputs for exercise Atbash Cipher - Goal: https://github.com/exercism/todo/issues/13
--------------
**Skillset:** This is mostly a matter of reading code (test suites) in several languages and creating some JSON.
---------------
Create a file `atbash-cipher.json` containing the inputs and expected outputs for Atbash Cipher.
Atbash Cipher has been implemented in the following languages:
- https://github.com/exercism/xclojure/tree/master/atbash-cipher
- https://github.com/exercism/xcoffeescript/tree/master/atbash-cipher
- https://github.com/exercism/xcsharp/tree/master/atbash-cipher
- https://github.com/exercism/xelixir/tree/master/atbash-cipher
- https://github.com/exercism/xerlang/tree/master/atbash-cipher
- https://github.com/exercism/xfsharp/tree/master/atbash-cipher
- https://github.com/exercism/xgo/tree/master/atbash-cipher
- https://github.com/exercism/xhaskell/tree/master/atbash-cipher
- https://github.com/exercism/xjava/tree/master/atbash-cipher
- https://github.com/exercism/xjavascript/tree/master/atbash-cipher
- https://github.com/exercism/xlfe/tree/master/atbash-cipher
- https://github.com/exercism/xlisp/tree/master/atbash-cipher
- https://github.com/exercism/xperl5/tree/master/atbash-cipher
- https://github.com/exercism/xpython/tree/master/atbash-cipher
- https://github.com/exercism/xruby/tree/master/atbash-cipher
- https://github.com/exercism/xscala/tree/master/atbash-cipher
- https://github.com/exercism/xswift/tree/master/atbash-cipher
See the following files for some examples:
- https://github.com/exercism/x-common/blob/master/bob.json
- https://github.com/exercism/x-common/blob/master/clock.json
- https://github.com/exercism/x-common/blob/master/custom-set.json
- https://github.com/exercism/x-common/blob/master/gigasecond.json
- https://github.com/exercism/x-common/blob/master/hamming.json
- https://github.com/exercism/x-common/blob/master/leap.json
|
non_build
|
extract shared inputs outputs for exercise atbash cipher goal skillset this is mostly a matter of reading code test suites in several languages and creating some json create a file atbash cipher json containing the inputs and expected outputs for atbash cipher atbash cipher has been implemented in the following languages see the following files for some examples
| 0
|
20,570
| 6,899,635,415
|
IssuesEvent
|
2017-11-24 14:37:25
|
csmk/frabjous
|
https://api.github.com/repos/csmk/frabjous
|
closed
|
net-im/profanity: new package (v0.5.1)
|
in progress new ebuild
|
> _**Profanity** is a console based XMPP client inspired by [Irssi](http://www.irssi.org/)_
- http://profanity.im
- https://github.com/boothj5/profanity
|
1.0
|
net-im/profanity: new package (v0.5.1) - > _**Profanity** is a console based XMPP client inspired by [Irssi](http://www.irssi.org/)_
- http://profanity.im
- https://github.com/boothj5/profanity
|
build
|
net im profanity new package profanity is a console based xmpp client inspired by
| 1
|
27,057
| 4,866,161,034
|
IssuesEvent
|
2016-11-14 22:50:31
|
arescentral/antares
|
https://api.github.com/repos/arescentral/antares
|
closed
|
.gn missing from 0.8.1 source release
|
Complexity:Low Priority:Critical Type:Defect
|
It's needed to build Antares, but it's missing from the source release because scripts/dist.py excludes all dotfiles.
|
1.0
|
.gn missing from 0.8.1 source release - It's needed to build Antares, but it's missing from the source release because scripts/dist.py excludes all dotfiles.
|
non_build
|
gn missing from source release it s needed to build antares but it s missing from the source release because scripts dist py excludes all dotfiles
| 0
|
152,061
| 5,832,225,330
|
IssuesEvent
|
2017-05-08 21:16:42
|
vmware/vic
|
https://api.github.com/repos/vmware/vic
|
closed
|
write unit tests to validate api input data for UpdateCluster
|
area/kubernetes priority/medium product/vcs
|
*User Statement*
As a VIC Admin, I expect the api to act appropriately to invalid input data for the UpdateCluster endpoint
*Config schema*
```yaml
updateCluster:
type: object
required:
- name
- minNodes
- noOfMasters
properties:
name:
type: string
description: the cluster name, should be valid for use in dns names
readOnly: true
credentials:
$ref: "#/definitions/credentials"
thumbprint:
description: the thumbprint of the vCenter server certificate
type: string
minLength: 57
pattern: '[a-fA-F0-9:]+'
minNodes:
type: integer
format: int32
min: 1
description: the minimum number of nodes that can be deployed
maxNodes:
type: integer
format: int32
description: the minimum number of nodes that can be deployed
noOfMasters:
type: integer
format: int32
description: the number of master nodes to create
default: 1
storageClasses:
type: array
items:
$ref: "#/definitions/storageClass"
nodeResourcePools:
type: array
items:
type: string
minLength: 1
```
*Acceptance Criteria*
unit tests that validate input data
Every validation needs to be covered
Also cover the valid case
|
1.0
|
write unit tests to validate api input data for UpdateCluster - *User Statement*
As a VIC Admin, I expect the api to act appropriately to invalid input data for the UpdateCluster endpoint
*Config schema*
```yaml
updateCluster:
type: object
required:
- name
- minNodes
- noOfMasters
properties:
name:
type: string
description: the cluster name, should be valid for use in dns names
readOnly: true
credentials:
$ref: "#/definitions/credentials"
thumbprint:
description: the thumbprint of the vCenter server certificate
type: string
minLength: 57
pattern: '[a-fA-F0-9:]+'
minNodes:
type: integer
format: int32
min: 1
description: the minimum number of nodes that can be deployed
maxNodes:
type: integer
format: int32
description: the minimum number of nodes that can be deployed
noOfMasters:
type: integer
format: int32
description: the number of master nodes to create
default: 1
storageClasses:
type: array
items:
$ref: "#/definitions/storageClass"
nodeResourcePools:
type: array
items:
type: string
minLength: 1
```
*Acceptance Criteria*
unit tests that validate input data
Every validation needs to be covered
Also cover the valid case
|
non_build
|
write unit tests to validate api input data for updatecluster user statement as a vic admin i expect the api to act appropriately to invalid input data for the updatecluster endpoint config schema yaml updatecluster type object required name minnodes noofmasters properties name type string description the cluster name should be valid for use in dns names readonly true credentials ref definitions credentials thumbprint description the thumbprint of the vcenter server certificate type string minlength pattern minnodes type integer format min description the minimum number of nodes that can be deployed maxnodes type integer format description the minimum number of nodes that can be deployed noofmasters type integer format description the number of master nodes to create default storageclasses type array items ref definitions storageclass noderesourcepools type array items type string minlength acceptance criteria unit tests that validate input data every validation needs to be covered also cover the valid case
| 0
|
4,034
| 2,812,082,970
|
IssuesEvent
|
2015-05-18 05:30:26
|
Mikeware/SpaceBattleArena
|
https://api.github.com/repos/Mikeware/SpaceBattleArena
|
closed
|
Fix website sidebar navigation from covering text.
|
bug documentation enhancement in progress
|
Thought I had tested this, should either leave fixed or left-alight content and right-align navigation maybe?
|
1.0
|
Fix website sidebar navigation from covering text. - Thought I had tested this, should either leave fixed or left-alight content and right-align navigation maybe?
|
non_build
|
fix website sidebar navigation from covering text thought i had tested this should either leave fixed or left alight content and right align navigation maybe
| 0
|
46,763
| 13,181,695,571
|
IssuesEvent
|
2020-08-12 14:41:53
|
xinYG/bootstrap-timepicker
|
https://api.github.com/repos/xinYG/bootstrap-timepicker
|
opened
|
CVE-2018-1000620 (High) detected in cryptiles-2.0.5.tgz
|
security vulnerability
|
## CVE-2018-1000620 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary>
<p>General purpose crypto utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/bootstrap-timepicker/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/bootstrap-timepicker/node_modules/cryptiles/package.json</p>
<p>
Dependency Hierarchy:
- grunt-contrib-jasmine-0.4.2.tgz (Root Library)
- grunt-lib-phantomjs-0.3.1.tgz
- phantomjs-1.9.20.tgz
- request-2.67.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/xinYG/bootstrap-timepicker/commit/c92f0918f68b35842a2bf5ae212e5d75e70546e5">c92f0918f68b35842a2bf5ae212e5d75e70546e5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p>
<p>Release Date: 2018-07-09</p>
<p>Fix Resolution: v4.1.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"cryptiles","packageVersion":"2.0.5","isTransitiveDependency":true,"dependencyTree":"grunt-contrib-jasmine:0.4.2;grunt-lib-phantomjs:0.3.1;phantomjs:1.9.20;request:2.67.0;hawk:3.1.3;cryptiles:2.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.1.2"}],"vulnerabilityIdentifier":"CVE-2018-1000620","vulnerabilityDetails":"Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2018-1000620 (High) detected in cryptiles-2.0.5.tgz - ## CVE-2018-1000620 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary>
<p>General purpose crypto utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/bootstrap-timepicker/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/bootstrap-timepicker/node_modules/cryptiles/package.json</p>
<p>
Dependency Hierarchy:
- grunt-contrib-jasmine-0.4.2.tgz (Root Library)
- grunt-lib-phantomjs-0.3.1.tgz
- phantomjs-1.9.20.tgz
- request-2.67.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/xinYG/bootstrap-timepicker/commit/c92f0918f68b35842a2bf5ae212e5d75e70546e5">c92f0918f68b35842a2bf5ae212e5d75e70546e5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p>
<p>Release Date: 2018-07-09</p>
<p>Fix Resolution: v4.1.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"cryptiles","packageVersion":"2.0.5","isTransitiveDependency":true,"dependencyTree":"grunt-contrib-jasmine:0.4.2;grunt-lib-phantomjs:0.3.1;phantomjs:1.9.20;request:2.67.0;hawk:3.1.3;cryptiles:2.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.1.2"}],"vulnerabilityIdentifier":"CVE-2018-1000620","vulnerabilityDetails":"Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_build
|
cve high detected in cryptiles tgz cve high severity vulnerability vulnerable library cryptiles tgz general purpose crypto utilities library home page a href path to dependency file tmp ws scm bootstrap timepicker package json path to vulnerable library tmp ws scm bootstrap timepicker node modules cryptiles package json dependency hierarchy grunt contrib jasmine tgz root library grunt lib phantomjs tgz phantomjs tgz request tgz hawk tgz x cryptiles tgz vulnerable library found in head commit a href vulnerability details eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in vulnerabilityurl
| 0
|
54,855
| 13,461,155,871
|
IssuesEvent
|
2020-09-09 14:28:59
|
NVIDIA/spark-rapids
|
https://api.github.com/repos/NVIDIA/spark-rapids
|
closed
|
[BUG] udf compiler should be part of the dist jar
|
P1 build
|
We should make the UDF compiler be a part of the dist jar, else people can't get to it for release 0.2.
For now, this can work like the shuffle plugin where it gets added to the dist jar. But discussing with @jlowe, he argued we can flatten the module structure and have everything in one module, maybe that second part can be done in release 0.3. I'll file something different for that.
|
1.0
|
[BUG] udf compiler should be part of the dist jar - We should make the UDF compiler be a part of the dist jar, else people can't get to it for release 0.2.
For now, this can work like the shuffle plugin where it gets added to the dist jar. But discussing with @jlowe, he argued we can flatten the module structure and have everything in one module, maybe that second part can be done in release 0.3. I'll file something different for that.
|
build
|
udf compiler should be part of the dist jar we should make the udf compiler be a part of the dist jar else people can t get to it for release for now this can work like the shuffle plugin where it gets added to the dist jar but discussing with jlowe he argued we can flatten the module structure and have everything in one module maybe that second part can be done in release i ll file something different for that
| 1
|
583,982
| 17,402,384,902
|
IssuesEvent
|
2021-08-02 21:49:05
|
certbot/certbot
|
https://api.github.com/repos/certbot/certbot
|
closed
|
Windows download
|
area: windows priority: critical
|
Windows 10
I Haven't been able to install cert bot as the windows download link seems to have expired!
|
1.0
|
Windows download - Windows 10
I Haven't been able to install cert bot as the windows download link seems to have expired!
|
non_build
|
windows download windows i haven t been able to install cert bot as the windows download link seems to have expired
| 0
|
58,065
| 14,271,907,388
|
IssuesEvent
|
2020-11-21 14:41:56
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
closed
|
So frustrating with this error: Failed to load the native TensorFlow runtime
|
TF 2.0 stalled stat:awaiting response subtype:macOS type:build/install
|
<em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): OS X
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): pip install tensorflow
- TensorFlow version: 2.0.1
- Python version: 3.7.3
- Installed using virtualenv? pip? conda?: pip
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version:
- GPU model and memory:
**Describe the problem**
It is so frustrating that when i google this error, so many webpages are found but no (working) solutions are provided. If this is such a common error, can you guys provide a fix or an official solution for fixing the problem?
Basically, i am trying to install tensorflow on my Mac. So what i did is:
pip install tensorflow
It was finished without any error.
But when I try:
import tensorflow as tf
The common import error came up: Failed to load the native TensorFlow runtime.
I couldn't figure out what is the exact issue, is it because Python 3.7 does not work with tensorflow 2.0.1? I tried:
pip install tensorflow==1.15.0.
But it says no 1.15.0 found in 2.0.1, not sure what does that mean.
**Provide the exact sequence of commands / steps that you executed before running into the problem**
pip install tensorflow
import tensorflow as tf
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
1.0
|
So frustrating with this error: Failed to load the native TensorFlow runtime - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): OS X
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): pip install tensorflow
- TensorFlow version: 2.0.1
- Python version: 3.7.3
- Installed using virtualenv? pip? conda?: pip
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version:
- GPU model and memory:
**Describe the problem**
It is so frustrating that when i google this error, so many webpages are found but no (working) solutions are provided. If this is such a common error, can you guys provide a fix or an official solution for fixing the problem?
Basically, i am trying to install tensorflow on my Mac. So what i did is:
pip install tensorflow
It was finished without any error.
But when I try:
import tensorflow as tf
The common import error came up: Failed to load the native TensorFlow runtime.
I couldn't figure out what is the exact issue, is it because Python 3.7 does not work with tensorflow 2.0.1? I tried:
pip install tensorflow==1.15.0.
But it says no 1.15.0 found in 2.0.1, not sure what does that mean.
**Provide the exact sequence of commands / steps that you executed before running into the problem**
pip install tensorflow
import tensorflow as tf
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
build
|
so frustrating with this error failed to load the native tensorflow runtime please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu os x mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary pip install tensorflow tensorflow version python version installed using virtualenv pip conda pip bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory describe the problem it is so frustrating that when i google this error so many webpages are found but no working solutions are provided if this is such a common error can you guys provide a fix or an official solution for fixing the problem basically i am trying to install tensorflow on my mac so what i did is pip install tensorflow it was finished without any error but when i try import tensorflow as tf the common import error came up failed to load the native tensorflow runtime i couldn t figure out what is the exact issue is it because python does not work with tensorflow i tried pip install tensorflow but it says no found in not sure what does that mean provide the exact sequence of commands steps that you executed before running into the problem pip install tensorflow import tensorflow as tf any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
| 1
|
58,999
| 14,522,374,483
|
IssuesEvent
|
2020-12-14 08:44:11
|
datainsider-co/rocket-bi-web
|
https://api.github.com/repos/datainsider-co/rocket-bi-web
|
closed
|
[Config panel]: Drop area wrong design
|
bug data builder
|
Sai UI ở các điểm:
1. Kích thước block [Drag& drop] ở config to hơn so với ở filter.
2. Text [Drag& drop] ở config không đúng vị trí center.

|
1.0
|
[Config panel]: Drop area wrong design - Sai UI ở các điểm:
1. Kích thước block [Drag& drop] ở config to hơn so với ở filter.
2. Text [Drag& drop] ở config không đúng vị trí center.

|
build
|
drop area wrong design sai ui ở các điểm kích thước block ở config to hơn so với ở filter text ở config không đúng vị trí center
| 1
|
60,825
| 25,275,335,470
|
IssuesEvent
|
2022-11-16 12:14:46
|
serie-a-logistics-solutions/nx3_demo_status
|
https://api.github.com/repos/serie-a-logistics-solutions/nx3_demo_status
|
opened
|
🛑 Database Service is down
|
status database-service
|
In [`0bf5f5f`](https://github.com/serie-a-logistics-solutions/nx3_demo_status/commit/0bf5f5f3c3b6a0181411dfa051b1c063c2e63ec1
), Database Service (https://transport-demo.on.nx3.cloud/monitoring/node1/rs/actuator/health) was **down**:
- HTTP code: 0
- Response time: 0 ms
|
1.0
|
🛑 Database Service is down - In [`0bf5f5f`](https://github.com/serie-a-logistics-solutions/nx3_demo_status/commit/0bf5f5f3c3b6a0181411dfa051b1c063c2e63ec1
), Database Service (https://transport-demo.on.nx3.cloud/monitoring/node1/rs/actuator/health) was **down**:
- HTTP code: 0
- Response time: 0 ms
|
non_build
|
🛑 database service is down in database service was down http code response time ms
| 0
|
23,095
| 10,851,570,281
|
IssuesEvent
|
2019-11-13 11:01:14
|
Sh2dowFi3nd/test
|
https://api.github.com/repos/Sh2dowFi3nd/test
|
opened
|
CVE-2016-6802 (High) detected in shiro-web-1.2.0.jar
|
security vulnerability
|
## CVE-2016-6802 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shiro-web-1.2.0.jar</b></p></summary>
<p>Apache Shiro is a powerful and flexible open-source security framework that cleanly handles
authentication, authorization, enterprise session management, single sign-on and cryptography services.</p>
<p>Library home page: <a href="http://shiro.apache.org/shiro-web/">http://shiro.apache.org/shiro-web/</a></p>
<p>
Dependency Hierarchy:
- :x: **shiro-web-1.2.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Sh2dowFi3nd/test/commit/5981a223918363e27b537d0da5292bdb1b6e70cf">5981a223918363e27b537d0da5292bdb1b6e70cf</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Shiro before 1.3.2 allows attackers to bypass intended servlet filters and gain access by leveraging use of a non-root servlet context path.
<p>Publish Date: 2016-09-20
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6802>CVE-2016-6802</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6802">https://nvd.nist.gov/vuln/detail/CVE-2016-6802</a></p>
<p>Release Date: 2016-09-20</p>
<p>Fix Resolution: 1.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-6802 (High) detected in shiro-web-1.2.0.jar - ## CVE-2016-6802 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shiro-web-1.2.0.jar</b></p></summary>
<p>Apache Shiro is a powerful and flexible open-source security framework that cleanly handles
authentication, authorization, enterprise session management, single sign-on and cryptography services.</p>
<p>Library home page: <a href="http://shiro.apache.org/shiro-web/">http://shiro.apache.org/shiro-web/</a></p>
<p>
Dependency Hierarchy:
- :x: **shiro-web-1.2.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Sh2dowFi3nd/test/commit/5981a223918363e27b537d0da5292bdb1b6e70cf">5981a223918363e27b537d0da5292bdb1b6e70cf</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Shiro before 1.3.2 allows attackers to bypass intended servlet filters and gain access by leveraging use of a non-root servlet context path.
<p>Publish Date: 2016-09-20
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6802>CVE-2016-6802</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6802">https://nvd.nist.gov/vuln/detail/CVE-2016-6802</a></p>
<p>Release Date: 2016-09-20</p>
<p>Fix Resolution: 1.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_build
|
cve high detected in shiro web jar cve high severity vulnerability vulnerable library shiro web jar apache shiro is a powerful and flexible open source security framework that cleanly handles authentication authorization enterprise session management single sign on and cryptography services library home page a href dependency hierarchy x shiro web jar vulnerable library found in head commit a href vulnerability details apache shiro before allows attackers to bypass intended servlet filters and gain access by leveraging use of a non root servlet context path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
16,719
| 6,265,993,970
|
IssuesEvent
|
2017-07-16 22:06:29
|
zfsonlinux/zfs
|
https://api.github.com/repos/zfsonlinux/zfs
|
closed
|
Failed for build latest ZoL's master branch update as buitin on the zen-kernel 4.3/master & same error with vanilla kernel linux-4.3
|
Build Issue
|
LD [M] drivers/staging/lustre/lustre/ptlrpc/ptlrpc.o
Building modules, stage 2.
MODPOST 3987 modules
ERROR: "spl_kmem_alloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_panic" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_vmem_zalloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_vmem_alloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_vmem_free" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_kmem_zalloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_kmem_zalloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kmem_asprintf" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_member" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_set_raw_ops" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_waitq_exit" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetfsuid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_alloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_cancel_id" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "z_compress_level" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_free" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "z_uncompress" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_openat" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_wait_outstanding" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetfsgid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__thread_exit" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "system_taskq" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_broadcast" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_runq_exit" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__udivdi3" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetgid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_releasef" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vcmn_err" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kmem_vasprintf" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_vmem_zalloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetuid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_wait_io" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_panic" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "ddi_strtoull" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_rdwr" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_signal" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_remove" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "cmn_err" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_get" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_install" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_init_ent" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kmem_debugging" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_dispatch" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "ddi_copyout" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_wait_sig" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_set" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "strfree" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "zone_get_hostid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_vmem_alloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_getattr" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_waitq_enter" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_wait" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_fsync" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_runq_enter" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_get_filesize" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_getf" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "groupmember" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_open_file" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_open" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetngroups" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "heap_arena" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetruid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetgroups" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_delete" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_set_pwd" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crfree" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "p0" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "strdup" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_read_file" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_close_file" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "rootdir" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vmem_size" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__umoddi3" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_wait" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_mode_to_vtype" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_timedwait_sig" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_alloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_seek" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_dispatch_ent" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "cv_timedwait_hires" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "ddi_copyin" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__thread_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_wait_id" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_set_move" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crhold" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_reap_now" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_close" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "zio_arena" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_init" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_vmem_free" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__divdi3" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_dispatch_delay" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_alloc" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "__umoddi3" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "cmn_err" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "spl_kmem_alloc" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "spl_panic" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "spl_vmem_alloc" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "ddi_strtol" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "xdrmem_create" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/nvpair/znvpair.ko] undefined!
WARNING: modpost: Found 5 section mismatch(es).
To see full details build your kernel with:
'make CONFIG_DEBUG_SECTION_MISMATCH=y'
scripts/Makefile.modpost:90: recipe for target '__modpost' failed
make[2]: **\* [__modpost] Error 1
Makefile:1095: error make recipe for target «modules»
make[1]: **\* [modules] Error 2
make[1]: exit dir «/media/BUILD/zen++»
debian/ruleset/targets/common.mk:295: error for make target «debian/stamp/build/kernel»
make: **\* [debian/stamp/build/kernel] Error 2
spl & zfs master branches was cloning and configured with:
./autogen.sh
./configure --enable-linux-builtin --with-linux=../../zen++ (my local build kernel dir)
./copy-builtin ../../zen++
for eatch.
The spl was change as builtin [*] by menuconfig and zfs was sort as module [M]
Stage 1 and 2 of kernel build process leave without errors, but this is trap!
Any solutions?Ideas?Sorry for my bad english...
|
1.0
|
Failed for build latest ZoL's master branch update as buitin on the zen-kernel 4.3/master & same error with vanilla kernel linux-4.3 - LD [M] drivers/staging/lustre/lustre/ptlrpc/ptlrpc.o
Building modules, stage 2.
MODPOST 3987 modules
ERROR: "spl_kmem_alloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_panic" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_vmem_zalloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_vmem_alloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_vmem_free" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_kmem_zalloc" [fs/zfs/zpios/zpios.ko] undefined!
ERROR: "spl_kmem_zalloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kmem_asprintf" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_member" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_set_raw_ops" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_waitq_exit" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetfsuid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_alloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_cancel_id" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "z_compress_level" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_free" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "z_uncompress" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_openat" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_wait_outstanding" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetfsgid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__thread_exit" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "system_taskq" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_broadcast" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_runq_exit" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__udivdi3" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetgid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_releasef" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vcmn_err" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kmem_vasprintf" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_vmem_zalloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetuid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_wait_io" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_panic" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "ddi_strtoull" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_rdwr" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_signal" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_remove" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "cmn_err" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_get" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_install" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_init_ent" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kmem_debugging" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_dispatch" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "ddi_copyout" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_wait_sig" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "tsd_set" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "strfree" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "zone_get_hostid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_vmem_alloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_getattr" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_waitq_enter" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_wait" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_fsync" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kstat_runq_enter" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_get_filesize" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_getf" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "groupmember" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_open_file" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_open" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetngroups" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "heap_arena" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetruid" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crgetgroups" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__kstat_delete" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_set_pwd" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crfree" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "p0" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "strdup" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_read_file" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "kobj_close_file" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "rootdir" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vmem_size" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__umoddi3" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_wait" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_mode_to_vtype" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_timedwait_sig" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_alloc" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_destroy" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_seek" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_dispatch_ent" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "cv_timedwait_hires" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "ddi_copyin" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__thread_create" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_wait_id" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_set_move" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "crhold" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_cache_reap_now" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "vn_close" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "zio_arena" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__cv_init" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_vmem_free" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "__divdi3" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "taskq_dispatch_delay" [fs/zfs/zfs/zfs.ko] undefined!
ERROR: "spl_kmem_alloc" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "__umoddi3" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "cmn_err" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/zcommon/zcommon.ko] undefined!
ERROR: "spl_kmem_alloc" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "spl_panic" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "spl_vmem_alloc" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "ddi_strtol" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "xdrmem_create" [fs/zfs/nvpair/znvpair.ko] undefined!
ERROR: "spl_kmem_free" [fs/zfs/nvpair/znvpair.ko] undefined!
WARNING: modpost: Found 5 section mismatch(es).
To see full details build your kernel with:
'make CONFIG_DEBUG_SECTION_MISMATCH=y'
scripts/Makefile.modpost:90: recipe for target '__modpost' failed
make[2]: **\* [__modpost] Error 1
Makefile:1095: error make recipe for target «modules»
make[1]: **\* [modules] Error 2
make[1]: exit dir «/media/BUILD/zen++»
debian/ruleset/targets/common.mk:295: error for make target «debian/stamp/build/kernel»
make: **\* [debian/stamp/build/kernel] Error 2
spl & zfs master branches was cloning and configured with:
./autogen.sh
./configure --enable-linux-builtin --with-linux=../../zen++ (my local build kernel dir)
./copy-builtin ../../zen++
for eatch.
The spl was change as builtin [*] by menuconfig and zfs was sort as module [M]
Stage 1 and 2 of kernel build process leave without errors, but this is trap!
Any solutions?Ideas?Sorry for my bad english...
|
build
|
failed for build latest zol s master branch update as buitin on the zen kernel master same error with vanilla kernel linux ld drivers staging lustre lustre ptlrpc ptlrpc o building modules stage modpost modules error spl kmem alloc undefined error spl panic undefined error spl vmem zalloc undefined error spl vmem alloc undefined error spl kmem free undefined error spl vmem free undefined error spl kmem zalloc undefined error spl kmem zalloc undefined error kmem asprintf undefined error taskq member undefined error spl kmem cache destroy undefined error kstat set raw ops undefined error kstat waitq exit undefined error crgetfsuid undefined error spl kmem cache alloc undefined error taskq cancel id undefined error z compress level undefined error spl kmem cache free undefined error z uncompress undefined error vn openat undefined error taskq wait outstanding undefined error crgetfsgid undefined error tsd create undefined error thread exit undefined error system taskq undefined error cv broadcast undefined error kstat create undefined error kstat runq exit undefined error undefined error taskq destroy undefined error crgetgid undefined error vn releasef undefined error vcmn err undefined error kmem vasprintf undefined error spl vmem zalloc undefined error crgetuid undefined error cv wait io undefined error spl panic undefined error tsd destroy undefined error ddi strtoull undefined error vn rdwr undefined error cv signal undefined error vn remove undefined error cmn err undefined error tsd get undefined error spl kmem free undefined error kstat install undefined error taskq init ent undefined error kmem debugging undefined error taskq dispatch undefined error ddi copyout undefined error cv wait sig undefined error tsd set undefined error strfree undefined error zone get hostid undefined error spl vmem alloc undefined error vn getattr undefined error kstat waitq enter undefined error cv wait undefined error vn fsync undefined error kstat runq enter undefined error kobj get filesize undefined error vn getf undefined error groupmember undefined error kobj open file undefined error vn open undefined error crgetngroups undefined error heap arena undefined error crgetruid undefined error crgetgroups undefined error taskq create undefined error kstat delete undefined error vn set pwd undefined error crfree undefined error undefined error strdup undefined error kobj read file undefined error kobj close file undefined error rootdir undefined error vmem size undefined error undefined error spl kmem cache create undefined error taskq wait undefined error vn mode to vtype undefined error cv timedwait sig undefined error spl kmem alloc undefined error cv destroy undefined error vn seek undefined error taskq dispatch ent undefined error cv timedwait hires undefined error ddi copyin undefined error thread create undefined error taskq wait id undefined error spl kmem cache set move undefined error crhold undefined error spl kmem cache reap now undefined error vn close undefined error zio arena undefined error cv init undefined error spl vmem free undefined error undefined error taskq dispatch delay undefined error spl kmem alloc undefined error undefined error cmn err undefined error spl kmem free undefined error spl kmem alloc undefined error spl panic undefined error spl vmem alloc undefined error ddi strtol undefined error xdrmem create undefined error spl kmem free undefined warning modpost found section mismatch es to see full details build your kernel with make config debug section mismatch y scripts makefile modpost recipe for target modpost failed make error makefile error make recipe for target «modules» make error make exit dir « media build zen » debian ruleset targets common mk error for make target «debian stamp build kernel» make error spl zfs master branches was cloning and configured with autogen sh configure enable linux builtin with linux zen my local build kernel dir copy builtin zen for eatch the spl was change as builtin by menuconfig and zfs was sort as module stage and of kernel build process leave without errors but this is trap any solutions ideas sorry for my bad english
| 1
|
277,354
| 24,063,590,751
|
IssuesEvent
|
2022-09-17 06:30:01
|
freqtrade/freqtrade
|
https://api.github.com/repos/freqtrade/freqtrade
|
closed
|
Backtesting results and hyperopt produce different outcomes
|
Question Backtest Strategy assistance
|
## Describe your environment
* Operating system: Windows 10 (Ubuntu version: 2004.2022.1.0)
* Python Version: 3.10.5
* CCXT version: 1.92.52
* Freqtrade Version: 2022.8
## Describe the problem:
I have a problem mentioned in the title, but I just cannot figure out why. At first I thout hat maybe theres an issue with QTpyLib Bollinger Bands, but it's not. It looks like the [issue in this topic](https://github.com/freqtrade/freqtrade/issues/7173), but it's actually not. I left a message there, but the issue is closed.
Here's an example of what's happening:
```
PS C:\Freqtrade\ft_userdata> docker-compose run freqtrade hyperopt --hyperopt-loss SortinoHyperOptLoss -s broken_result --spaces buy, roi, stoploss, trailing -e 1000 -i 1h --timerange 20210422-20220908
Creating ft_userdata_freqtrade_run ... done
2022-09-16 12:55:02,883 - freqtrade.configuration.load_config - INFO - Using config: user_data/config.json ...
2022-09-16 12:55:02,885 - freqtrade.loggers - INFO - Verbosity set to 0
2022-09-16 12:55:02,886 - freqtrade.configuration.configuration - INFO - Parameter -i/--timeframe detected ... Using timeframe: 1h ...
2022-09-16 12:55:02,886 - freqtrade.configuration.configuration - INFO - Using max_open_trades: 80 ...
2022-09-16 12:55:02,886 - freqtrade.configuration.configuration - INFO - Parameter --timerange detected: 20210422-20220908 ...
2022-09-16 12:55:04,353 - freqtrade.configuration.configuration - INFO - Using user-data directory: /freqtrade/user_data ...
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Using data directory: /freqtrade/user_data/data/binance ...
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Overriding timeframe with Command line argument
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter --epochs detected ... Will run Hyperopt with for 1000 epochs ...
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter -s/--spaces detected: ['buy', 'roi', 'stoploss', 'trailing']
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter -j/--job-workers detected: -1
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter --min-trades detected: 1
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Using Hyperopt loss class name: SortinoHyperOptLoss
2022-09-16 12:55:04,354 - freqtrade.configuration.check_exchange - INFO - Checking exchange...
2022-09-16 12:55:04,365 - freqtrade.configuration.check_exchange - INFO - Exchange "binance" is officially supported by the Freqtrade development team.
2022-09-16 12:55:04,365 - freqtrade.configuration.configuration - INFO - Using pairlist from configuration.
2022-09-16 12:55:04,366 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:55:04,368 - freqtrade.commands.optimize_commands - INFO - Starting freqtrade in Hyperopt mode
2022-09-16 12:55:04,370 - freqtrade.exchange.exchange - INFO - Instance is running with dry_run enabled
2022-09-16 12:55:04,370 - freqtrade.exchange.exchange - INFO - Using CCXT 1.92.52
2022-09-16 12:55:04,387 - freqtrade.exchange.exchange - INFO - Using Exchange "Binance"
2022-09-16 12:55:06,024 - freqtrade.resolvers.exchange_resolver - INFO - Using resolved exchange 'Binance'...
2022-09-16 12:55:06,105 - freqtrade.resolvers.iresolver - INFO - Using resolved strategy broken_result from '/freqtrade/user_data/strategies/broken_result.py'...
2022-09-16 12:55:06,106 - freqtrade.strategy.hyper - INFO - Loading parameters from file /freqtrade/user_data/strategies/broken_result.json
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'timeframe' with value in config file: 1h.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_currency' with value in config file: USDT.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_amount' with value in config file: 100.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'unfilledtimeout' with value in config file: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using minimal_roi: {'0': 0.375, '471': 0.089, '675': 0.033, '1123': 0}
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using timeframe: 1h
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stoploss: -0.302
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop: True
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive: 0.014
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive_offset: 0.079
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_only_offset_is_reached: False
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_custom_stoploss: False
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using process_only_new_candles: True
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_types: {'entry': 'limit', 'exit': 'limit', 'stoploss': 'limit', 'stoploss_on_exchange': False, 'stoploss_on_exchange_interval': 60}
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_time_in_force: {'entry': 'gtc', 'exit': 'gtc'}
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_currency: USDT
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_amount: 100
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using protections: []
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using startup_candle_count: 0
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using unfilledtimeout: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_exit_signal: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_only: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_roi_if_entry_signal: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_offset: 0.0
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_buying_expired_candle_after: 0
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using position_adjustment_enable: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using max_entry_position_adjustment: -1
2022-09-16 12:55:06,110 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:55:06,115 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist StaticPairList from '/freqtrade/freqtrade/plugins/pairlist/StaticPairList.py'...
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - Strategy Parameter: bb_time = 47
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - Strategy Parameter: roc = 1.97
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - No params for sell found, using default values.
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - No params for protection found, using default values.
2022-09-16 12:55:06,195 - freqtrade.resolvers.iresolver - INFO - Using resolved hyperoptloss SortinoHyperOptLoss from '/freqtrade/freqtrade/optimize/hyperopt_loss/hyperopt_loss_sortino.py'...
2022-09-16 12:55:06,196 - freqtrade.optimize.hyperopt - INFO - Removing `/freqtrade/user_data/hyperopt_results/hyperopt_tickerdata.pkl`.
2022-09-16 12:55:06,201 - freqtrade.optimize.hyperopt - INFO - Using optimizer random state: 42402
2022-09-16 12:55:06,203 - freqtrade.optimize.hyperopt_interface - INFO - Min roi table: {0: 0.069, 120: 0.046, 240: 0.023, 360: 0}
2022-09-16 12:55:06,203 - freqtrade.optimize.hyperopt_interface - INFO - Max roi table: {0: 0.711, 480: 0.252, 1200: 0.092, 2640: 0}
2022-09-16 12:55:06,700 - freqtrade.data.history.idatahandler - WARNING - SHIB/USDT, spot, 1h, data starts at 2021-05-10 11:00:00
2022-09-16 12:55:07,141 - freqtrade.data.history.idatahandler - WARNING - FLOW/USDT, spot, 1h, data starts at 2021-07-30 13:00:00
2022-09-16 12:55:07,230 - freqtrade.data.history.idatahandler - WARNING - ICP/USDT, spot, 1h, data starts at 2021-05-11 01:00:00
2022-09-16 12:55:07,607 - freqtrade.data.history.idatahandler - WARNING - QNT/USDT, spot, 1h, data starts at 2021-07-29 06:00:00
2022-09-16 12:55:07,733 - freqtrade.optimize.backtesting - INFO - Loading data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days).
2022-09-16 12:55:07,733 - freqtrade.optimize.hyperopt - INFO - Dataload complete. Calculating indicators
2022-09-16 12:55:07,925 - freqtrade.optimize.hyperopt - INFO - Hyperopting with data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days)..
2022-09-16 12:55:08,207 - freqtrade.exchange.exchange - INFO - Closing async ccxt session.
2022-09-16 12:55:08,220 - freqtrade.optimize.hyperopt - INFO - Found 6 CPU cores. Let's make them scream!
2022-09-16 12:55:08,220 - freqtrade.optimize.hyperopt - INFO - Number of parallel jobs set as: -1
2022-09-16 12:55:08,220 - freqtrade.optimize.hyperopt - INFO - Using estimator ET.
2022-09-16 12:55:08,229 - freqtrade.optimize.hyperopt - INFO - Effective number of parallel workers used: 6
+--------+-----------+----------+------------------+--------------+-------------------------------+-----------------+-------------+-------------------------------+
| Best | Epoch | Trades | Win Draw Loss | Avg profit | Profit | Avg duration | Objective | Max Drawdown (Acct) |
|--------+-----------+----------+------------------+--------------+-------------------------------+-----------------+-------------+-------------------------------|
| * Best | 1/1000 | 373 | 188 0 185 | -1.99% | -742.811 USDT (-74.28%) | 0 days 20:35:00 | 6.84283 | 762.562 USDT (76.00%) |
| * Best | 3/1000 | 190 | 124 0 66 | -2.03% | -385.886 USDT (-38.59%) | 0 days 22:06:00 | 2.22487 | 482.823 USDT (46.35%) |
| * Best | 4/1000 | 239 | 201 0 38 | -1.17% | -279.853 USDT (-27.99%) | 1 days 09:43:00 | 1.13748 | 419.488 USDT (37.97%) |
| * Best | 9/1000 | 144 | 117 0 27 | -1.72% | -246.833 USDT (-24.68%) | 1 days 04:31:00 | 1.02734 | 322.905 USDT (31.14%) |
| * Best | 10/1000 | 211 | 178 0 33 | -0.85% | -178.724 USDT (-17.87%) | 1 days 04:18:00 | 0.837 | 290.169 USDT (27.43%) |
| Best | 33/1000 | 174 | 145 0 29 | -0.47% | -81.804 USDT (-8.18%) | 1 days 05:56:00 | 0.40621 | 212.048 USDT (19.63%) |
[Epoch 37 of 1000 ( 3%)] |███████| | [ETA: 0:36:57, Elapsed Time: 0:01:25]^C
User interrupted..
2022-09-16 12:56:49,530 - freqtrade.optimize.hyperopt - INFO - 42 epochs saved to '/freqtrade/user_data/hyperopt_results/strategy_broken_result_2022-09-16_12-55-06.fthypt'.
2022-09-16 12:56:49,744 - freqtrade.resolvers.iresolver - WARNING - Could not import /freqtrade/user_data/strategies/KSC_close.py due to 'invalid syntax (KSC_close.py, line 167)'
2022-09-16 12:56:50,032 - freqtrade.optimize.hyperopt_tools - INFO - Dumping parameters to /freqtrade/user_data/strategies/broken_result.json
Best result:
33/1000: 174 trades. 145/0/29 Wins/Draws/Losses. Avg profit -0.47%. Median profit 3.30%. Total profit -81.80389263 USDT ( -8.18%). Avg duration 1 day, 5:56:00 min. Objective: 0.40621
# Buy hyperspace params:
buy_params = {
"bb_time": 39,
"roc": 1.84,
}
# ROI table:
minimal_roi = {
"0": 0.544,
"173": 0.121,
"753": 0.033,
"2036": 0
}
# Stoploss:
stoploss = -0.226
# Trailing stop:
trailing_stop = True
trailing_stop_positive = 0.159
trailing_stop_positive_offset = 0.172
trailing_only_offset_is_reached = True
PS C:\Freqtrade\ft_userdata> docker-compose run freqtrade backtesting --datadir user_data/data/binance --export trades --stake-amount 100 -s broken_result -i 1h --timerange 20210422-20220908
Creating ft_userdata_freqtrade_run ... done
2022-09-16 12:57:07,080 - freqtrade.configuration.load_config - INFO - Using config: user_data/config.json ...
2022-09-16 12:57:07,082 - freqtrade.loggers - INFO - Verbosity set to 0
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Parameter -i/--timeframe detected ... Using timeframe: 1h ...
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Using max_open_trades: 80 ...
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Parameter --stake-amount detected, overriding stake_amount to: 100.0 ...
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Parameter --timerange detected: 20210422-20220908 ...
2022-09-16 12:57:08,444 - freqtrade.configuration.configuration - INFO - Using user-data directory: /freqtrade/user_data ...
2022-09-16 12:57:08,445 - freqtrade.configuration.configuration - INFO - Using data directory: user_data/data/binance ...
2022-09-16 12:57:08,445 - freqtrade.configuration.configuration - INFO - Overriding timeframe with Command line argument
2022-09-16 12:57:08,445 - freqtrade.configuration.configuration - INFO - Parameter --export detected: trades ...
2022-09-16 12:57:08,446 - freqtrade.configuration.configuration - INFO - Parameter --cache=day detected ...
2022-09-16 12:57:08,446 - freqtrade.configuration.check_exchange - INFO - Checking exchange...
2022-09-16 12:57:08,457 - freqtrade.configuration.check_exchange - INFO - Exchange "binance" is officially supported by the Freqtrade development team.
2022-09-16 12:57:08,457 - freqtrade.configuration.configuration - INFO - Using pairlist from configuration.
2022-09-16 12:57:08,457 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:57:08,460 - freqtrade.commands.optimize_commands - INFO - Starting freqtrade in Backtesting mode
2022-09-16 12:57:08,461 - freqtrade.exchange.exchange - INFO - Instance is running with dry_run enabled
2022-09-16 12:57:08,461 - freqtrade.exchange.exchange - INFO - Using CCXT 1.92.52
2022-09-16 12:57:08,477 - freqtrade.exchange.exchange - INFO - Using Exchange "Binance"
2022-09-16 12:57:10,172 - freqtrade.resolvers.exchange_resolver - INFO - Using resolved exchange 'Binance'...
2022-09-16 12:57:10,249 - freqtrade.resolvers.iresolver - INFO - Using resolved strategy broken_result from '/freqtrade/user_data/strategies/broken_result.py'...
2022-09-16 12:57:10,250 - freqtrade.strategy.hyper - INFO - Loading parameters from file /freqtrade/user_data/strategies/broken_result.json
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'timeframe' with value in config file: 1h.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_currency' with value in config file: USDT.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_amount' with value in config file: 100.0.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'unfilledtimeout' with value in config file: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using minimal_roi: {'0': 0.544, '173': 0.121, '753': 0.033, '2036': 0}
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using timeframe: 1h
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stoploss: -0.226
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop: True
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive: 0.159
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive_offset: 0.172
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_only_offset_is_reached: True
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_custom_stoploss: False
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using process_only_new_candles: True
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_types: {'entry': 'limit', 'exit': 'limit', 'stoploss': 'limit', 'stoploss_on_exchange': False, 'stoploss_on_exchange_interval': 60}
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_time_in_force: {'entry': 'gtc', 'exit': 'gtc'}
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_currency: USDT
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_amount: 100.0
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using protections: []
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using startup_candle_count: 0
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using unfilledtimeout: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_exit_signal: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_only: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_roi_if_entry_signal: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_offset: 0.0
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using disable_dataframe_checks: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_buying_expired_candle_after: 0
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using position_adjustment_enable: False
2022-09-16 12:57:10,256 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using max_entry_position_adjustment: -1
2022-09-16 12:57:10,256 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:57:10,260 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist StaticPairList from '/freqtrade/freqtrade/plugins/pairlist/StaticPairList.py'...
2022-09-16 12:57:10,810 - freqtrade.data.history.idatahandler - WARNING - SHIB/USDT, spot, 1h, data starts at 2021-05-10 11:00:00
2022-09-16 12:57:11,239 - freqtrade.data.history.idatahandler - WARNING - FLOW/USDT, spot, 1h, data starts at 2021-07-30 13:00:00
2022-09-16 12:57:11,329 - freqtrade.data.history.idatahandler - WARNING - ICP/USDT, spot, 1h, data starts at 2021-05-11 01:00:00
2022-09-16 12:57:11,784 - freqtrade.data.history.idatahandler - WARNING - QNT/USDT, spot, 1h, data starts at 2021-07-29 06:00:00
2022-09-16 12:57:11,830 - freqtrade.optimize.backtesting - INFO - Loading data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days).
2022-09-16 12:57:11,830 - freqtrade.optimize.backtesting - INFO - Dataload complete. Calculating indicators
2022-09-16 12:57:11,857 - freqtrade.optimize.backtesting - INFO - Running backtesting for Strategy broken_result
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - Strategy Parameter: bb_time = 39
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - Strategy Parameter: roc = 1.84
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - No params for sell found, using default values.
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - No params for protection found, using default values.
2022-09-16 12:57:12,052 - freqtrade.optimize.backtesting - INFO - Backtesting with data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days).
2022-09-16 12:57:14,883 - freqtrade.misc - INFO - dumping json to "/freqtrade/user_data/backtest_results/backtest-result-2022-09-16_12-57-14.meta.json"
2022-09-16 12:57:14,885 - freqtrade.misc - INFO - dumping json to "/freqtrade/user_data/backtest_results/backtest-result-2022-09-16_12-57-14.json"
2022-09-16 12:57:14,891 - freqtrade.misc - INFO - dumping json to "/freqtrade/user_data/backtest_results/.last_result.json"
Result for strategy broken_result
============================================================ BACKTESTING REPORT ===========================================================
| Pair | Buys | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % | Avg Duration | Win Draw Loss Win% |
|------------+--------+----------------+----------------+-------------------+----------------+------------------+-------------------------|
| UNI/USDT | 5 | 4.69 | 23.45 | 23.458 | 2.35 | 22:36:00 | 5 0 0 100 |
| FIL/USDT | 6 | 2.81 | 16.87 | 16.817 | 1.68 | 21:30:00 | 6 0 0 100 |
| HNT/USDT | 4 | 4.02 | 16.08 | 16.075 | 1.61 | 14:30:00 | 4 0 0 100 |
| MANA/USDT | 3 | 5.15 | 15.44 | 15.444 | 1.54 | 23:20:00 | 3 0 0 100 |
| AXS/USDT | 5 | 2.68 | 13.39 | 13.378 | 1.34 | 1 day, 10:24:00 | 5 0 0 100 |
| LTC/USDT | 6 | 2.09 | 12.55 | 12.556 | 1.26 | 1 day, 9:30:00 | 6 0 0 100 |
| SAND/USDT | 1 | 12.12 | 12.12 | 12.102 | 1.21 | 3:00:00 | 1 0 0 100 |
| ETH/USDT | 1 | 12.09 | 12.09 | 12.089 | 1.21 | 7:00:00 | 1 0 0 100 |
| HBAR/USDT | 2 | 3.32 | 6.64 | 6.637 | 0.66 | 16:00:00 | 2 0 0 100 |
| BCH/USDT | 4 | 1.66 | 6.64 | 6.631 | 0.66 | 1 day, 10:00:00 | 4 0 0 100 |
| XRP/USDT | 3 | 2.21 | 6.62 | 6.619 | 0.66 | 1 day, 3:20:00 | 3 0 0 100 |
| AVAX/USDT | 4 | 1.20 | 4.81 | 4.793 | 0.48 | 13:45:00 | 3 0 1 75.0 |
| QNT/USDT | 1 | 3.35 | 3.35 | 3.355 | 0.34 | 20:00:00 | 1 0 0 100 |
| EOS/USDT | 2 | 1.66 | 3.32 | 3.311 | 0.33 | 23:30:00 | 2 0 0 100 |
| ICP/USDT | 6 | 0.08 | 0.50 | 0.455 | 0.05 | 1 day, 19:20:00 | 5 0 1 83.3 |
| XTZ/USDT | 2 | 0.02 | 0.04 | 0.037 | 0.00 | 1 day, 17:00:00 | 2 0 0 100 |
| BTC/USDT | 1 | 0.00 | 0.00 | 0.000 | 0.00 | 1 day, 10:00:00 | 1 0 0 100 |
| THETA/USDT | 4 | -1.63 | -6.51 | -6.425 | -0.64 | 23:30:00 | 3 0 1 75.0 |
| DOT/USDT | 5 | -1.44 | -7.22 | -7.248 | -0.72 | 21:12:00 | 4 0 1 80.0 |
| LINK/USDT | 4 | -1.84 | -7.35 | -7.348 | -0.73 | 1 day, 6:30:00 | 3 0 1 75.0 |
| SHIB/USDT | 5 | -2.31 | -11.56 | -11.570 | -1.16 | 13:24:00 | 4 0 1 80.0 |
| EGLD/USDT | 6 | -2.14 | -12.84 | -12.817 | -1.28 | 1 day, 19:00:00 | 5 0 1 83.3 |
| ALGO/USDT | 4 | -3.20 | -12.81 | -12.860 | -1.29 | 14:00:00 | 3 0 1 75.0 |
| XMR/USDT | 4 | -4.02 | -16.07 | -16.026 | -1.60 | 1 day, 11:00:00 | 3 0 1 75.0 |
| ETC/USDT | 6 | -2.68 | -16.11 | -16.058 | -1.61 | 1 day, 11:50:00 | 5 0 1 83.3 |
| FLOW/USDT | 4 | -4.02 | -16.08 | -16.086 | -1.61 | 2 days, 4:00:00 | 3 0 1 75.0 |
| FTT/USDT | 4 | -4.39 | -17.56 | -17.539 | -1.75 | 1 day, 7:15:00 | 3 0 1 75.0 |
| VET/USDT | 4 | -4.86 | -19.43 | -19.446 | -1.94 | 1 day, 10:30:00 | 3 0 1 75.0 |
| ATOM/USDT | 2 | -9.72 | -19.44 | -19.456 | -1.95 | 19:00:00 | 1 0 1 50.0 |
| NEAR/USDT | 1 | -22.75 | -22.75 | -22.731 | -2.27 | 20:00:00 | 0 0 1 0 |
| TRX/USDT | 1 | -22.75 | -22.75 | -22.772 | -2.28 | 1 day, 18:00:00 | 0 0 1 0 |
| SOL/USDT | 7 | -3.35 | -23.46 | -23.196 | -2.32 | 1 day, 6:17:00 | 5 0 2 71.4 |
| AAVE/USDT | 4 | -7.50 | -29.99 | -29.901 | -2.99 | 14:30:00 | 2 0 2 50.0 |
| MATIC/USDT | 6 | -5.38 | -32.31 | -32.324 | -3.23 | 13:30:00 | 4 0 2 66.7 |
| XLM/USDT | 3 | -11.13 | -33.39 | -33.340 | -3.33 | 21:40:00 | 1 0 2 33.3 |
| DOGE/USDT | 4 | -8.57 | -34.28 | -34.319 | -3.43 | 19:00:00 | 2 0 2 50.0 |
| ADA/USDT | 6 | -6.33 | -37.99 | -37.985 | -3.80 | 3 days, 13:20:00 | 4 0 2 66.7 |
| TOTAL | 140 | -1.76 | -246.00 | -245.692 | -24.57 | 1 day, 5:32:00 | 112 0 28 80.0 |
========================================================= ENTER TAG STATS ==========================================================
| TAG | Buys | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % | Avg Duration | Win Draw Loss Win% |
|-------+--------+----------------+----------------+-------------------+----------------+----------------+-------------------------|
| TOTAL | 140 | -1.76 | -246.00 | -245.692 | -24.57 | 1 day, 5:32:00 | 112 0 28 80.0 |
======================================================= EXIT REASON STATS ========================================================
| Exit Reason | Exits | Win Draws Loss Win% | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % |
|--------------------+---------+--------------------------+----------------+----------------+-------------------+----------------|
| roi | 109 | 109 0 0 100 | 3.44 | 375.04 | 374.791 | 10.14 |
| stop_loss | 28 | 0 0 28 0 | -22.74 | -636.82 | -636.235 | -17.21 |
| trailing_stop_loss | 3 | 3 0 0 100 | 5.26 | 15.78 | 15.752 | 0.43 |
====================================================== LEFT OPEN TRADES REPORT ======================================================
| Pair | Buys | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % | Avg Duration | Win Draw Loss Win% |
|--------+--------+----------------+----------------+-------------------+----------------+----------------+-------------------------|
| TOTAL | 0 | 0.00 | 0.00 | 0.000 | 0.00 | 0:00 | 0 0 0 0 |
================== SUMMARY METRICS ==================
| Metric | Value |
|-----------------------------+---------------------|
| Backtesting from | 2021-04-22 00:00:00 |
| Backtesting to | 2022-09-08 00:00:00 |
| Max open trades | 37 |
| | |
| Total/Daily Avg Trades | 140 / 0.28 |
| Starting balance | 1000 USDT |
| Final balance | 754.308 USDT |
| Absolute profit | -245.692 USDT |
| Total profit % | -24.57% |
| CAGR % | -18.47% |
| Profit factor | 0.61 |
| Trades per day | 0.28 |
| Avg. daily profit % | -0.05% |
| Avg. stake amount | 99.799 USDT |
| Total trade volume | 13971.907 USDT |
| | |
| Best Pair | UNI/USDT 23.45% |
| Worst Pair | ADA/USDT -37.99% |
| Best trade | UNI/USDT 16.66% |
| Worst trade | ADA/USDT -22.75% |
| Best day | 25.091 USDT |
| Worst day | -146.95 USDT |
| Days win/draw/lose | 59 / 434 / 10 |
| Avg. Duration Winners | 1 day, 2:36:00 |
| Avg. Duration Loser | 1 day, 17:15:00 |
| Rejected Entry signals | 0 |
| Entry/Exit Timeouts | 2 / 0 |
| | |
| Min balance | 708.658 USDT |
| Max balance | 1053.937 USDT |
| Max % of account underwater | 32.76% |
| Absolute Drawdown (Account) | 32.76% |
| Absolute Drawdown | 345.279 USDT |
| Drawdown high | 53.937 USDT |
| Drawdown low | -291.342 USDT |
| Drawdown Start | 2021-05-22 20:00:00 |
| Drawdown End | 2022-05-12 07:00:00 |
| Market change | -50.31% |
=====================================================
2022-09-16 12:57:14,975 - freqtrade.exchange.exchange - INFO - Closing async ccxt session.
```
There's no changes between hyperopt and backtest, same config file, but results are different o_0
Example of the strategy where this happens is below (I know it's bad, but that's just an example):
```
# --- Do not remove these libs ---
import numpy as np # noqa
import pandas as pd # noqa
import datetime
from pandas import DataFrame
from freqtrade.strategy import (BooleanParameter, CategoricalParameter, DecimalParameter,
IStrategy, IntParameter, merge_informative_pair)
# --------------------------------
# Add your lib to import here
import talib
import talib.abstract as ta
import pandas_ta as pta
import freqtrade.vendor.qtpylib.indicators as qtpylib
from technical.util import resample_to_interval, resampled_merge
class broken_result(IStrategy):
# ROI table:
minimal_roi = {
"0": 0.62,
"456": 0.211,
"737": 0.051,
"2120": 0
}
# Stoploss:
stoploss = -0.026
# Trailing stop:
trailing_stop = True
trailing_stop_positive = 0.234
trailing_stop_positive_offset = 0.292
trailing_only_offset_is_reached = True
#exit_profit_only = True
use_exit_signal = False
# Optimal timeframe for the strategy
timeframe = '1h'
roc = DecimalParameter(low=1, high=2, decimals=2, default=1.38, space="buy", optimize=True)
bb_time = IntParameter(low=14, high=50, default=20, space='buy', optimize=True) #40
def populate_indicators(self, dataframe: DataFrame, metadata: dict) -> DataFrame:
################################
# # Stochastic Slow
stoch = ta.STOCH(dataframe, 14, 3, 0, 3, 0)
dataframe['slowd'] = stoch['slowd']
dataframe['slowk'] = stoch['slowk']
dataframe['roc'] = ta.ROC(dataframe, timeperiod=1)
bollinger = qtpylib.bollinger_bands(dataframe['close'], window=self.bb_time.value, stds=2)
dataframe['bb_lowerband'] = bollinger['lower']
dataframe['bb_upperband'] = bollinger['upper']
dataframe['bb_middleband'] = bollinger['mid']
# dataframe['bb_middleband'] = dataframe['close'].rolling(self.bb_time.value).mean()
# rolling_std = dataframe['close'].rolling(self.bb_time.value).std()
# dataframe['bb_upperband'] = dataframe['bb_middleband'] + (2 * rolling_std)
# dataframe['bb_lowerband'] = dataframe['bb_middleband'] - (2 * rolling_std)
dataframe['volume_max'] = ta.MAX(dataframe['volume'], timeperiod=10) #10
return dataframe
def populate_buy_trend(self, dataframe: DataFrame, metadata: dict) -> DataFrame:
dataframe.loc[
(
(dataframe['close'] > dataframe['close'].shift(1)) &
(dataframe['roc'] > self.roc.value) &
(qtpylib.crossed_above(dataframe['slowk'], dataframe['slowd'])) &
(dataframe['volume'] < dataframe['volume_max'].shift(1)) &
(dataframe['close'] < dataframe['bb_lowerband'])
)
,
'buy'] = 1
return dataframe
def populate_sell_trend(self, dataframe: DataFrame, metadata: dict) -> DataFrame:
"""
Based on TA indicators, populates the sell signal for the given dataframe
:param dataframe: DataFrame
:return: DataFrame with buy column
"""
dataframe.loc[
(
)
,
'sell'
] = 0
return dataframe
```
### Steps to reproduce:
1. Run Hyperopt for a few epochs and stop it (or let it finish).
2. Run Backtest (using generated .json file or copying hyperopt parameters into the strategy)
3. Observe
### Observed Results:
* What happened? - Results are different
* What did you expect to happen? - They should be the same
### Notes
I think hyperopt has a problem with optimizing the timeperiod.
|
1.0
|
Backtesting results and hyperopt produce different outcomes - ## Describe your environment
* Operating system: Windows 10 (Ubuntu version: 2004.2022.1.0)
* Python Version: 3.10.5
* CCXT version: 1.92.52
* Freqtrade Version: 2022.8
## Describe the problem:
I have a problem mentioned in the title, but I just cannot figure out why. At first I thout hat maybe theres an issue with QTpyLib Bollinger Bands, but it's not. It looks like the [issue in this topic](https://github.com/freqtrade/freqtrade/issues/7173), but it's actually not. I left a message there, but the issue is closed.
Here's an example of what's happening:
```
PS C:\Freqtrade\ft_userdata> docker-compose run freqtrade hyperopt --hyperopt-loss SortinoHyperOptLoss -s broken_result --spaces buy, roi, stoploss, trailing -e 1000 -i 1h --timerange 20210422-20220908
Creating ft_userdata_freqtrade_run ... done
2022-09-16 12:55:02,883 - freqtrade.configuration.load_config - INFO - Using config: user_data/config.json ...
2022-09-16 12:55:02,885 - freqtrade.loggers - INFO - Verbosity set to 0
2022-09-16 12:55:02,886 - freqtrade.configuration.configuration - INFO - Parameter -i/--timeframe detected ... Using timeframe: 1h ...
2022-09-16 12:55:02,886 - freqtrade.configuration.configuration - INFO - Using max_open_trades: 80 ...
2022-09-16 12:55:02,886 - freqtrade.configuration.configuration - INFO - Parameter --timerange detected: 20210422-20220908 ...
2022-09-16 12:55:04,353 - freqtrade.configuration.configuration - INFO - Using user-data directory: /freqtrade/user_data ...
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Using data directory: /freqtrade/user_data/data/binance ...
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Overriding timeframe with Command line argument
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter --epochs detected ... Will run Hyperopt with for 1000 epochs ...
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter -s/--spaces detected: ['buy', 'roi', 'stoploss', 'trailing']
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter -j/--job-workers detected: -1
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Parameter --min-trades detected: 1
2022-09-16 12:55:04,354 - freqtrade.configuration.configuration - INFO - Using Hyperopt loss class name: SortinoHyperOptLoss
2022-09-16 12:55:04,354 - freqtrade.configuration.check_exchange - INFO - Checking exchange...
2022-09-16 12:55:04,365 - freqtrade.configuration.check_exchange - INFO - Exchange "binance" is officially supported by the Freqtrade development team.
2022-09-16 12:55:04,365 - freqtrade.configuration.configuration - INFO - Using pairlist from configuration.
2022-09-16 12:55:04,366 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:55:04,368 - freqtrade.commands.optimize_commands - INFO - Starting freqtrade in Hyperopt mode
2022-09-16 12:55:04,370 - freqtrade.exchange.exchange - INFO - Instance is running with dry_run enabled
2022-09-16 12:55:04,370 - freqtrade.exchange.exchange - INFO - Using CCXT 1.92.52
2022-09-16 12:55:04,387 - freqtrade.exchange.exchange - INFO - Using Exchange "Binance"
2022-09-16 12:55:06,024 - freqtrade.resolvers.exchange_resolver - INFO - Using resolved exchange 'Binance'...
2022-09-16 12:55:06,105 - freqtrade.resolvers.iresolver - INFO - Using resolved strategy broken_result from '/freqtrade/user_data/strategies/broken_result.py'...
2022-09-16 12:55:06,106 - freqtrade.strategy.hyper - INFO - Loading parameters from file /freqtrade/user_data/strategies/broken_result.json
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'timeframe' with value in config file: 1h.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_currency' with value in config file: USDT.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_amount' with value in config file: 100.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'unfilledtimeout' with value in config file: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}.
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using minimal_roi: {'0': 0.375, '471': 0.089, '675': 0.033, '1123': 0}
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using timeframe: 1h
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stoploss: -0.302
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop: True
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive: 0.014
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive_offset: 0.079
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_only_offset_is_reached: False
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_custom_stoploss: False
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using process_only_new_candles: True
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_types: {'entry': 'limit', 'exit': 'limit', 'stoploss': 'limit', 'stoploss_on_exchange': False, 'stoploss_on_exchange_interval': 60}
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_time_in_force: {'entry': 'gtc', 'exit': 'gtc'}
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_currency: USDT
2022-09-16 12:55:06,109 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_amount: 100
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using protections: []
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using startup_candle_count: 0
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using unfilledtimeout: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_exit_signal: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_only: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_roi_if_entry_signal: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_offset: 0.0
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_buying_expired_candle_after: 0
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using position_adjustment_enable: False
2022-09-16 12:55:06,110 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using max_entry_position_adjustment: -1
2022-09-16 12:55:06,110 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:55:06,115 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist StaticPairList from '/freqtrade/freqtrade/plugins/pairlist/StaticPairList.py'...
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - Strategy Parameter: bb_time = 47
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - Strategy Parameter: roc = 1.97
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - No params for sell found, using default values.
2022-09-16 12:55:06,184 - freqtrade.strategy.hyper - INFO - No params for protection found, using default values.
2022-09-16 12:55:06,195 - freqtrade.resolvers.iresolver - INFO - Using resolved hyperoptloss SortinoHyperOptLoss from '/freqtrade/freqtrade/optimize/hyperopt_loss/hyperopt_loss_sortino.py'...
2022-09-16 12:55:06,196 - freqtrade.optimize.hyperopt - INFO - Removing `/freqtrade/user_data/hyperopt_results/hyperopt_tickerdata.pkl`.
2022-09-16 12:55:06,201 - freqtrade.optimize.hyperopt - INFO - Using optimizer random state: 42402
2022-09-16 12:55:06,203 - freqtrade.optimize.hyperopt_interface - INFO - Min roi table: {0: 0.069, 120: 0.046, 240: 0.023, 360: 0}
2022-09-16 12:55:06,203 - freqtrade.optimize.hyperopt_interface - INFO - Max roi table: {0: 0.711, 480: 0.252, 1200: 0.092, 2640: 0}
2022-09-16 12:55:06,700 - freqtrade.data.history.idatahandler - WARNING - SHIB/USDT, spot, 1h, data starts at 2021-05-10 11:00:00
2022-09-16 12:55:07,141 - freqtrade.data.history.idatahandler - WARNING - FLOW/USDT, spot, 1h, data starts at 2021-07-30 13:00:00
2022-09-16 12:55:07,230 - freqtrade.data.history.idatahandler - WARNING - ICP/USDT, spot, 1h, data starts at 2021-05-11 01:00:00
2022-09-16 12:55:07,607 - freqtrade.data.history.idatahandler - WARNING - QNT/USDT, spot, 1h, data starts at 2021-07-29 06:00:00
2022-09-16 12:55:07,733 - freqtrade.optimize.backtesting - INFO - Loading data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days).
2022-09-16 12:55:07,733 - freqtrade.optimize.hyperopt - INFO - Dataload complete. Calculating indicators
2022-09-16 12:55:07,925 - freqtrade.optimize.hyperopt - INFO - Hyperopting with data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days)..
2022-09-16 12:55:08,207 - freqtrade.exchange.exchange - INFO - Closing async ccxt session.
2022-09-16 12:55:08,220 - freqtrade.optimize.hyperopt - INFO - Found 6 CPU cores. Let's make them scream!
2022-09-16 12:55:08,220 - freqtrade.optimize.hyperopt - INFO - Number of parallel jobs set as: -1
2022-09-16 12:55:08,220 - freqtrade.optimize.hyperopt - INFO - Using estimator ET.
2022-09-16 12:55:08,229 - freqtrade.optimize.hyperopt - INFO - Effective number of parallel workers used: 6
+--------+-----------+----------+------------------+--------------+-------------------------------+-----------------+-------------+-------------------------------+
| Best | Epoch | Trades | Win Draw Loss | Avg profit | Profit | Avg duration | Objective | Max Drawdown (Acct) |
|--------+-----------+----------+------------------+--------------+-------------------------------+-----------------+-------------+-------------------------------|
| * Best | 1/1000 | 373 | 188 0 185 | -1.99% | -742.811 USDT (-74.28%) | 0 days 20:35:00 | 6.84283 | 762.562 USDT (76.00%) |
| * Best | 3/1000 | 190 | 124 0 66 | -2.03% | -385.886 USDT (-38.59%) | 0 days 22:06:00 | 2.22487 | 482.823 USDT (46.35%) |
| * Best | 4/1000 | 239 | 201 0 38 | -1.17% | -279.853 USDT (-27.99%) | 1 days 09:43:00 | 1.13748 | 419.488 USDT (37.97%) |
| * Best | 9/1000 | 144 | 117 0 27 | -1.72% | -246.833 USDT (-24.68%) | 1 days 04:31:00 | 1.02734 | 322.905 USDT (31.14%) |
| * Best | 10/1000 | 211 | 178 0 33 | -0.85% | -178.724 USDT (-17.87%) | 1 days 04:18:00 | 0.837 | 290.169 USDT (27.43%) |
| Best | 33/1000 | 174 | 145 0 29 | -0.47% | -81.804 USDT (-8.18%) | 1 days 05:56:00 | 0.40621 | 212.048 USDT (19.63%) |
[Epoch 37 of 1000 ( 3%)] |███████| | [ETA: 0:36:57, Elapsed Time: 0:01:25]^C
User interrupted..
2022-09-16 12:56:49,530 - freqtrade.optimize.hyperopt - INFO - 42 epochs saved to '/freqtrade/user_data/hyperopt_results/strategy_broken_result_2022-09-16_12-55-06.fthypt'.
2022-09-16 12:56:49,744 - freqtrade.resolvers.iresolver - WARNING - Could not import /freqtrade/user_data/strategies/KSC_close.py due to 'invalid syntax (KSC_close.py, line 167)'
2022-09-16 12:56:50,032 - freqtrade.optimize.hyperopt_tools - INFO - Dumping parameters to /freqtrade/user_data/strategies/broken_result.json
Best result:
33/1000: 174 trades. 145/0/29 Wins/Draws/Losses. Avg profit -0.47%. Median profit 3.30%. Total profit -81.80389263 USDT ( -8.18%). Avg duration 1 day, 5:56:00 min. Objective: 0.40621
# Buy hyperspace params:
buy_params = {
"bb_time": 39,
"roc": 1.84,
}
# ROI table:
minimal_roi = {
"0": 0.544,
"173": 0.121,
"753": 0.033,
"2036": 0
}
# Stoploss:
stoploss = -0.226
# Trailing stop:
trailing_stop = True
trailing_stop_positive = 0.159
trailing_stop_positive_offset = 0.172
trailing_only_offset_is_reached = True
PS C:\Freqtrade\ft_userdata> docker-compose run freqtrade backtesting --datadir user_data/data/binance --export trades --stake-amount 100 -s broken_result -i 1h --timerange 20210422-20220908
Creating ft_userdata_freqtrade_run ... done
2022-09-16 12:57:07,080 - freqtrade.configuration.load_config - INFO - Using config: user_data/config.json ...
2022-09-16 12:57:07,082 - freqtrade.loggers - INFO - Verbosity set to 0
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Parameter -i/--timeframe detected ... Using timeframe: 1h ...
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Using max_open_trades: 80 ...
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Parameter --stake-amount detected, overriding stake_amount to: 100.0 ...
2022-09-16 12:57:07,082 - freqtrade.configuration.configuration - INFO - Parameter --timerange detected: 20210422-20220908 ...
2022-09-16 12:57:08,444 - freqtrade.configuration.configuration - INFO - Using user-data directory: /freqtrade/user_data ...
2022-09-16 12:57:08,445 - freqtrade.configuration.configuration - INFO - Using data directory: user_data/data/binance ...
2022-09-16 12:57:08,445 - freqtrade.configuration.configuration - INFO - Overriding timeframe with Command line argument
2022-09-16 12:57:08,445 - freqtrade.configuration.configuration - INFO - Parameter --export detected: trades ...
2022-09-16 12:57:08,446 - freqtrade.configuration.configuration - INFO - Parameter --cache=day detected ...
2022-09-16 12:57:08,446 - freqtrade.configuration.check_exchange - INFO - Checking exchange...
2022-09-16 12:57:08,457 - freqtrade.configuration.check_exchange - INFO - Exchange "binance" is officially supported by the Freqtrade development team.
2022-09-16 12:57:08,457 - freqtrade.configuration.configuration - INFO - Using pairlist from configuration.
2022-09-16 12:57:08,457 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:57:08,460 - freqtrade.commands.optimize_commands - INFO - Starting freqtrade in Backtesting mode
2022-09-16 12:57:08,461 - freqtrade.exchange.exchange - INFO - Instance is running with dry_run enabled
2022-09-16 12:57:08,461 - freqtrade.exchange.exchange - INFO - Using CCXT 1.92.52
2022-09-16 12:57:08,477 - freqtrade.exchange.exchange - INFO - Using Exchange "Binance"
2022-09-16 12:57:10,172 - freqtrade.resolvers.exchange_resolver - INFO - Using resolved exchange 'Binance'...
2022-09-16 12:57:10,249 - freqtrade.resolvers.iresolver - INFO - Using resolved strategy broken_result from '/freqtrade/user_data/strategies/broken_result.py'...
2022-09-16 12:57:10,250 - freqtrade.strategy.hyper - INFO - Loading parameters from file /freqtrade/user_data/strategies/broken_result.json
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'timeframe' with value in config file: 1h.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_currency' with value in config file: USDT.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_amount' with value in config file: 100.0.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'unfilledtimeout' with value in config file: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}.
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using minimal_roi: {'0': 0.544, '173': 0.121, '753': 0.033, '2036': 0}
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using timeframe: 1h
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stoploss: -0.226
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop: True
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive: 0.159
2022-09-16 12:57:10,253 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive_offset: 0.172
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_only_offset_is_reached: True
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_custom_stoploss: False
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using process_only_new_candles: True
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_types: {'entry': 'limit', 'exit': 'limit', 'stoploss': 'limit', 'stoploss_on_exchange': False, 'stoploss_on_exchange_interval': 60}
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_time_in_force: {'entry': 'gtc', 'exit': 'gtc'}
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_currency: USDT
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_amount: 100.0
2022-09-16 12:57:10,254 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using protections: []
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using startup_candle_count: 0
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using unfilledtimeout: {'entry': 5, 'exit': 5, 'exit_timeout_count': 0, 'unit': 'minutes'}
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_exit_signal: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_only: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_roi_if_entry_signal: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using exit_profit_offset: 0.0
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using disable_dataframe_checks: False
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_buying_expired_candle_after: 0
2022-09-16 12:57:10,255 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using position_adjustment_enable: False
2022-09-16 12:57:10,256 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using max_entry_position_adjustment: -1
2022-09-16 12:57:10,256 - freqtrade.configuration.config_validation - INFO - Validating configuration ...
2022-09-16 12:57:10,260 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist StaticPairList from '/freqtrade/freqtrade/plugins/pairlist/StaticPairList.py'...
2022-09-16 12:57:10,810 - freqtrade.data.history.idatahandler - WARNING - SHIB/USDT, spot, 1h, data starts at 2021-05-10 11:00:00
2022-09-16 12:57:11,239 - freqtrade.data.history.idatahandler - WARNING - FLOW/USDT, spot, 1h, data starts at 2021-07-30 13:00:00
2022-09-16 12:57:11,329 - freqtrade.data.history.idatahandler - WARNING - ICP/USDT, spot, 1h, data starts at 2021-05-11 01:00:00
2022-09-16 12:57:11,784 - freqtrade.data.history.idatahandler - WARNING - QNT/USDT, spot, 1h, data starts at 2021-07-29 06:00:00
2022-09-16 12:57:11,830 - freqtrade.optimize.backtesting - INFO - Loading data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days).
2022-09-16 12:57:11,830 - freqtrade.optimize.backtesting - INFO - Dataload complete. Calculating indicators
2022-09-16 12:57:11,857 - freqtrade.optimize.backtesting - INFO - Running backtesting for Strategy broken_result
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - Strategy Parameter: bb_time = 39
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - Strategy Parameter: roc = 1.84
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - No params for sell found, using default values.
2022-09-16 12:57:11,857 - freqtrade.strategy.hyper - INFO - No params for protection found, using default values.
2022-09-16 12:57:12,052 - freqtrade.optimize.backtesting - INFO - Backtesting with data from 2021-04-22 00:00:00 up to 2022-09-08 00:00:00 (504 days).
2022-09-16 12:57:14,883 - freqtrade.misc - INFO - dumping json to "/freqtrade/user_data/backtest_results/backtest-result-2022-09-16_12-57-14.meta.json"
2022-09-16 12:57:14,885 - freqtrade.misc - INFO - dumping json to "/freqtrade/user_data/backtest_results/backtest-result-2022-09-16_12-57-14.json"
2022-09-16 12:57:14,891 - freqtrade.misc - INFO - dumping json to "/freqtrade/user_data/backtest_results/.last_result.json"
Result for strategy broken_result
============================================================ BACKTESTING REPORT ===========================================================
| Pair | Buys | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % | Avg Duration | Win Draw Loss Win% |
|------------+--------+----------------+----------------+-------------------+----------------+------------------+-------------------------|
| UNI/USDT | 5 | 4.69 | 23.45 | 23.458 | 2.35 | 22:36:00 | 5 0 0 100 |
| FIL/USDT | 6 | 2.81 | 16.87 | 16.817 | 1.68 | 21:30:00 | 6 0 0 100 |
| HNT/USDT | 4 | 4.02 | 16.08 | 16.075 | 1.61 | 14:30:00 | 4 0 0 100 |
| MANA/USDT | 3 | 5.15 | 15.44 | 15.444 | 1.54 | 23:20:00 | 3 0 0 100 |
| AXS/USDT | 5 | 2.68 | 13.39 | 13.378 | 1.34 | 1 day, 10:24:00 | 5 0 0 100 |
| LTC/USDT | 6 | 2.09 | 12.55 | 12.556 | 1.26 | 1 day, 9:30:00 | 6 0 0 100 |
| SAND/USDT | 1 | 12.12 | 12.12 | 12.102 | 1.21 | 3:00:00 | 1 0 0 100 |
| ETH/USDT | 1 | 12.09 | 12.09 | 12.089 | 1.21 | 7:00:00 | 1 0 0 100 |
| HBAR/USDT | 2 | 3.32 | 6.64 | 6.637 | 0.66 | 16:00:00 | 2 0 0 100 |
| BCH/USDT | 4 | 1.66 | 6.64 | 6.631 | 0.66 | 1 day, 10:00:00 | 4 0 0 100 |
| XRP/USDT | 3 | 2.21 | 6.62 | 6.619 | 0.66 | 1 day, 3:20:00 | 3 0 0 100 |
| AVAX/USDT | 4 | 1.20 | 4.81 | 4.793 | 0.48 | 13:45:00 | 3 0 1 75.0 |
| QNT/USDT | 1 | 3.35 | 3.35 | 3.355 | 0.34 | 20:00:00 | 1 0 0 100 |
| EOS/USDT | 2 | 1.66 | 3.32 | 3.311 | 0.33 | 23:30:00 | 2 0 0 100 |
| ICP/USDT | 6 | 0.08 | 0.50 | 0.455 | 0.05 | 1 day, 19:20:00 | 5 0 1 83.3 |
| XTZ/USDT | 2 | 0.02 | 0.04 | 0.037 | 0.00 | 1 day, 17:00:00 | 2 0 0 100 |
| BTC/USDT | 1 | 0.00 | 0.00 | 0.000 | 0.00 | 1 day, 10:00:00 | 1 0 0 100 |
| THETA/USDT | 4 | -1.63 | -6.51 | -6.425 | -0.64 | 23:30:00 | 3 0 1 75.0 |
| DOT/USDT | 5 | -1.44 | -7.22 | -7.248 | -0.72 | 21:12:00 | 4 0 1 80.0 |
| LINK/USDT | 4 | -1.84 | -7.35 | -7.348 | -0.73 | 1 day, 6:30:00 | 3 0 1 75.0 |
| SHIB/USDT | 5 | -2.31 | -11.56 | -11.570 | -1.16 | 13:24:00 | 4 0 1 80.0 |
| EGLD/USDT | 6 | -2.14 | -12.84 | -12.817 | -1.28 | 1 day, 19:00:00 | 5 0 1 83.3 |
| ALGO/USDT | 4 | -3.20 | -12.81 | -12.860 | -1.29 | 14:00:00 | 3 0 1 75.0 |
| XMR/USDT | 4 | -4.02 | -16.07 | -16.026 | -1.60 | 1 day, 11:00:00 | 3 0 1 75.0 |
| ETC/USDT | 6 | -2.68 | -16.11 | -16.058 | -1.61 | 1 day, 11:50:00 | 5 0 1 83.3 |
| FLOW/USDT | 4 | -4.02 | -16.08 | -16.086 | -1.61 | 2 days, 4:00:00 | 3 0 1 75.0 |
| FTT/USDT | 4 | -4.39 | -17.56 | -17.539 | -1.75 | 1 day, 7:15:00 | 3 0 1 75.0 |
| VET/USDT | 4 | -4.86 | -19.43 | -19.446 | -1.94 | 1 day, 10:30:00 | 3 0 1 75.0 |
| ATOM/USDT | 2 | -9.72 | -19.44 | -19.456 | -1.95 | 19:00:00 | 1 0 1 50.0 |
| NEAR/USDT | 1 | -22.75 | -22.75 | -22.731 | -2.27 | 20:00:00 | 0 0 1 0 |
| TRX/USDT | 1 | -22.75 | -22.75 | -22.772 | -2.28 | 1 day, 18:00:00 | 0 0 1 0 |
| SOL/USDT | 7 | -3.35 | -23.46 | -23.196 | -2.32 | 1 day, 6:17:00 | 5 0 2 71.4 |
| AAVE/USDT | 4 | -7.50 | -29.99 | -29.901 | -2.99 | 14:30:00 | 2 0 2 50.0 |
| MATIC/USDT | 6 | -5.38 | -32.31 | -32.324 | -3.23 | 13:30:00 | 4 0 2 66.7 |
| XLM/USDT | 3 | -11.13 | -33.39 | -33.340 | -3.33 | 21:40:00 | 1 0 2 33.3 |
| DOGE/USDT | 4 | -8.57 | -34.28 | -34.319 | -3.43 | 19:00:00 | 2 0 2 50.0 |
| ADA/USDT | 6 | -6.33 | -37.99 | -37.985 | -3.80 | 3 days, 13:20:00 | 4 0 2 66.7 |
| TOTAL | 140 | -1.76 | -246.00 | -245.692 | -24.57 | 1 day, 5:32:00 | 112 0 28 80.0 |
========================================================= ENTER TAG STATS ==========================================================
| TAG | Buys | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % | Avg Duration | Win Draw Loss Win% |
|-------+--------+----------------+----------------+-------------------+----------------+----------------+-------------------------|
| TOTAL | 140 | -1.76 | -246.00 | -245.692 | -24.57 | 1 day, 5:32:00 | 112 0 28 80.0 |
======================================================= EXIT REASON STATS ========================================================
| Exit Reason | Exits | Win Draws Loss Win% | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % |
|--------------------+---------+--------------------------+----------------+----------------+-------------------+----------------|
| roi | 109 | 109 0 0 100 | 3.44 | 375.04 | 374.791 | 10.14 |
| stop_loss | 28 | 0 0 28 0 | -22.74 | -636.82 | -636.235 | -17.21 |
| trailing_stop_loss | 3 | 3 0 0 100 | 5.26 | 15.78 | 15.752 | 0.43 |
====================================================== LEFT OPEN TRADES REPORT ======================================================
| Pair | Buys | Avg Profit % | Cum Profit % | Tot Profit USDT | Tot Profit % | Avg Duration | Win Draw Loss Win% |
|--------+--------+----------------+----------------+-------------------+----------------+----------------+-------------------------|
| TOTAL | 0 | 0.00 | 0.00 | 0.000 | 0.00 | 0:00 | 0 0 0 0 |
================== SUMMARY METRICS ==================
| Metric | Value |
|-----------------------------+---------------------|
| Backtesting from | 2021-04-22 00:00:00 |
| Backtesting to | 2022-09-08 00:00:00 |
| Max open trades | 37 |
| | |
| Total/Daily Avg Trades | 140 / 0.28 |
| Starting balance | 1000 USDT |
| Final balance | 754.308 USDT |
| Absolute profit | -245.692 USDT |
| Total profit % | -24.57% |
| CAGR % | -18.47% |
| Profit factor | 0.61 |
| Trades per day | 0.28 |
| Avg. daily profit % | -0.05% |
| Avg. stake amount | 99.799 USDT |
| Total trade volume | 13971.907 USDT |
| | |
| Best Pair | UNI/USDT 23.45% |
| Worst Pair | ADA/USDT -37.99% |
| Best trade | UNI/USDT 16.66% |
| Worst trade | ADA/USDT -22.75% |
| Best day | 25.091 USDT |
| Worst day | -146.95 USDT |
| Days win/draw/lose | 59 / 434 / 10 |
| Avg. Duration Winners | 1 day, 2:36:00 |
| Avg. Duration Loser | 1 day, 17:15:00 |
| Rejected Entry signals | 0 |
| Entry/Exit Timeouts | 2 / 0 |
| | |
| Min balance | 708.658 USDT |
| Max balance | 1053.937 USDT |
| Max % of account underwater | 32.76% |
| Absolute Drawdown (Account) | 32.76% |
| Absolute Drawdown | 345.279 USDT |
| Drawdown high | 53.937 USDT |
| Drawdown low | -291.342 USDT |
| Drawdown Start | 2021-05-22 20:00:00 |
| Drawdown End | 2022-05-12 07:00:00 |
| Market change | -50.31% |
=====================================================
2022-09-16 12:57:14,975 - freqtrade.exchange.exchange - INFO - Closing async ccxt session.
```
There's no changes between hyperopt and backtest, same config file, but results are different o_0
Example of the strategy where this happens is below (I know it's bad, but that's just an example):
```
# --- Do not remove these libs ---
import numpy as np # noqa
import pandas as pd # noqa
import datetime
from pandas import DataFrame
from freqtrade.strategy import (BooleanParameter, CategoricalParameter, DecimalParameter,
IStrategy, IntParameter, merge_informative_pair)
# --------------------------------
# Add your lib to import here
import talib
import talib.abstract as ta
import pandas_ta as pta
import freqtrade.vendor.qtpylib.indicators as qtpylib
from technical.util import resample_to_interval, resampled_merge
class broken_result(IStrategy):
# ROI table:
minimal_roi = {
"0": 0.62,
"456": 0.211,
"737": 0.051,
"2120": 0
}
# Stoploss:
stoploss = -0.026
# Trailing stop:
trailing_stop = True
trailing_stop_positive = 0.234
trailing_stop_positive_offset = 0.292
trailing_only_offset_is_reached = True
#exit_profit_only = True
use_exit_signal = False
# Optimal timeframe for the strategy
timeframe = '1h'
roc = DecimalParameter(low=1, high=2, decimals=2, default=1.38, space="buy", optimize=True)
bb_time = IntParameter(low=14, high=50, default=20, space='buy', optimize=True) #40
def populate_indicators(self, dataframe: DataFrame, metadata: dict) -> DataFrame:
################################
# # Stochastic Slow
stoch = ta.STOCH(dataframe, 14, 3, 0, 3, 0)
dataframe['slowd'] = stoch['slowd']
dataframe['slowk'] = stoch['slowk']
dataframe['roc'] = ta.ROC(dataframe, timeperiod=1)
bollinger = qtpylib.bollinger_bands(dataframe['close'], window=self.bb_time.value, stds=2)
dataframe['bb_lowerband'] = bollinger['lower']
dataframe['bb_upperband'] = bollinger['upper']
dataframe['bb_middleband'] = bollinger['mid']
# dataframe['bb_middleband'] = dataframe['close'].rolling(self.bb_time.value).mean()
# rolling_std = dataframe['close'].rolling(self.bb_time.value).std()
# dataframe['bb_upperband'] = dataframe['bb_middleband'] + (2 * rolling_std)
# dataframe['bb_lowerband'] = dataframe['bb_middleband'] - (2 * rolling_std)
dataframe['volume_max'] = ta.MAX(dataframe['volume'], timeperiod=10) #10
return dataframe
def populate_buy_trend(self, dataframe: DataFrame, metadata: dict) -> DataFrame:
dataframe.loc[
(
(dataframe['close'] > dataframe['close'].shift(1)) &
(dataframe['roc'] > self.roc.value) &
(qtpylib.crossed_above(dataframe['slowk'], dataframe['slowd'])) &
(dataframe['volume'] < dataframe['volume_max'].shift(1)) &
(dataframe['close'] < dataframe['bb_lowerband'])
)
,
'buy'] = 1
return dataframe
def populate_sell_trend(self, dataframe: DataFrame, metadata: dict) -> DataFrame:
"""
Based on TA indicators, populates the sell signal for the given dataframe
:param dataframe: DataFrame
:return: DataFrame with buy column
"""
dataframe.loc[
(
)
,
'sell'
] = 0
return dataframe
```
### Steps to reproduce:
1. Run Hyperopt for a few epochs and stop it (or let it finish).
2. Run Backtest (using generated .json file or copying hyperopt parameters into the strategy)
3. Observe
### Observed Results:
* What happened? - Results are different
* What did you expect to happen? - They should be the same
### Notes
I think hyperopt has a problem with optimizing the timeperiod.
|
non_build
|
backtesting results and hyperopt produce different outcomes describe your environment operating system windows ubuntu version python version ccxt version freqtrade version describe the problem i have a problem mentioned in the title but i just cannot figure out why at first i thout hat maybe theres an issue with qtpylib bollinger bands but it s not it looks like the but it s actually not i left a message there but the issue is closed here s an example of what s happening ps c freqtrade ft userdata docker compose run freqtrade hyperopt hyperopt loss sortinohyperoptloss s broken result spaces buy roi stoploss trailing e i timerange creating ft userdata freqtrade run done freqtrade configuration load config info using config user data config json freqtrade loggers info verbosity set to freqtrade configuration configuration info parameter i timeframe detected using timeframe freqtrade configuration configuration info using max open trades freqtrade configuration configuration info parameter timerange detected freqtrade configuration configuration info using user data directory freqtrade user data freqtrade configuration configuration info using data directory freqtrade user data data binance freqtrade configuration configuration info overriding timeframe with command line argument freqtrade configuration configuration info parameter epochs detected will run hyperopt with for epochs freqtrade configuration configuration info parameter s spaces detected freqtrade configuration configuration info parameter j job workers detected freqtrade configuration configuration info parameter min trades detected freqtrade configuration configuration info using hyperopt loss class name sortinohyperoptloss freqtrade configuration check exchange info checking exchange freqtrade configuration check exchange info exchange binance is officially supported by the freqtrade development team freqtrade configuration configuration info using pairlist from configuration freqtrade configuration config validation info validating configuration freqtrade commands optimize commands info starting freqtrade in hyperopt mode freqtrade exchange exchange info instance is running with dry run enabled freqtrade exchange exchange info using ccxt freqtrade exchange exchange info using exchange binance freqtrade resolvers exchange resolver info using resolved exchange binance freqtrade resolvers iresolver info using resolved strategy broken result from freqtrade user data strategies broken result py freqtrade strategy hyper info loading parameters from file freqtrade user data strategies broken result json freqtrade resolvers strategy resolver info override strategy timeframe with value in config file freqtrade resolvers strategy resolver info override strategy stake currency with value in config file usdt freqtrade resolvers strategy resolver info override strategy stake amount with value in config file freqtrade resolvers strategy resolver info override strategy unfilledtimeout with value in config file entry exit exit timeout count unit minutes freqtrade resolvers strategy resolver info strategy using minimal roi freqtrade resolvers strategy resolver info strategy using timeframe freqtrade resolvers strategy resolver info strategy using stoploss freqtrade resolvers strategy resolver info strategy using trailing stop true freqtrade resolvers strategy resolver info strategy using trailing stop positive freqtrade resolvers strategy resolver info strategy using trailing stop positive offset freqtrade resolvers strategy resolver info strategy using trailing only offset is reached false freqtrade resolvers strategy resolver info strategy using use custom stoploss false freqtrade resolvers strategy resolver info strategy using process only new candles true freqtrade resolvers strategy resolver info strategy using order types entry limit exit limit stoploss limit stoploss on exchange false stoploss on exchange interval freqtrade resolvers strategy resolver info strategy using order time in force entry gtc exit gtc freqtrade resolvers strategy resolver info strategy using stake currency usdt freqtrade resolvers strategy resolver info strategy using stake amount freqtrade resolvers strategy resolver info strategy using protections freqtrade resolvers strategy resolver info strategy using startup candle count freqtrade resolvers strategy resolver info strategy using unfilledtimeout entry exit exit timeout count unit minutes freqtrade resolvers strategy resolver info strategy using use exit signal false freqtrade resolvers strategy resolver info strategy using exit profit only false freqtrade resolvers strategy resolver info strategy using ignore roi if entry signal false freqtrade resolvers strategy resolver info strategy using exit profit offset freqtrade resolvers strategy resolver info strategy using ignore buying expired candle after freqtrade resolvers strategy resolver info strategy using position adjustment enable false freqtrade resolvers strategy resolver info strategy using max entry position adjustment freqtrade configuration config validation info validating configuration freqtrade resolvers iresolver info using resolved pairlist staticpairlist from freqtrade freqtrade plugins pairlist staticpairlist py freqtrade strategy hyper info strategy parameter bb time freqtrade strategy hyper info strategy parameter roc freqtrade strategy hyper info no params for sell found using default values freqtrade strategy hyper info no params for protection found using default values freqtrade resolvers iresolver info using resolved hyperoptloss sortinohyperoptloss from freqtrade freqtrade optimize hyperopt loss hyperopt loss sortino py freqtrade optimize hyperopt info removing freqtrade user data hyperopt results hyperopt tickerdata pkl freqtrade optimize hyperopt info using optimizer random state freqtrade optimize hyperopt interface info min roi table freqtrade optimize hyperopt interface info max roi table freqtrade data history idatahandler warning shib usdt spot data starts at freqtrade data history idatahandler warning flow usdt spot data starts at freqtrade data history idatahandler warning icp usdt spot data starts at freqtrade data history idatahandler warning qnt usdt spot data starts at freqtrade optimize backtesting info loading data from up to days freqtrade optimize hyperopt info dataload complete calculating indicators freqtrade optimize hyperopt info hyperopting with data from up to days freqtrade exchange exchange info closing async ccxt session freqtrade optimize hyperopt info found cpu cores let s make them scream freqtrade optimize hyperopt info number of parallel jobs set as freqtrade optimize hyperopt info using estimator et freqtrade optimize hyperopt info effective number of parallel workers used best epoch trades win draw loss avg profit profit avg duration objective max drawdown acct best usdt days usdt best usdt days usdt best usdt days usdt best usdt days usdt best usdt days usdt best usdt days usdt ███████ c user interrupted freqtrade optimize hyperopt info epochs saved to freqtrade user data hyperopt results strategy broken result fthypt freqtrade resolvers iresolver warning could not import freqtrade user data strategies ksc close py due to invalid syntax ksc close py line freqtrade optimize hyperopt tools info dumping parameters to freqtrade user data strategies broken result json best result trades wins draws losses avg profit median profit total profit usdt avg duration day min objective buy hyperspace params buy params bb time roc roi table minimal roi stoploss stoploss trailing stop trailing stop true trailing stop positive trailing stop positive offset trailing only offset is reached true ps c freqtrade ft userdata docker compose run freqtrade backtesting datadir user data data binance export trades stake amount s broken result i timerange creating ft userdata freqtrade run done freqtrade configuration load config info using config user data config json freqtrade loggers info verbosity set to freqtrade configuration configuration info parameter i timeframe detected using timeframe freqtrade configuration configuration info using max open trades freqtrade configuration configuration info parameter stake amount detected overriding stake amount to freqtrade configuration configuration info parameter timerange detected freqtrade configuration configuration info using user data directory freqtrade user data freqtrade configuration configuration info using data directory user data data binance freqtrade configuration configuration info overriding timeframe with command line argument freqtrade configuration configuration info parameter export detected trades freqtrade configuration configuration info parameter cache day detected freqtrade configuration check exchange info checking exchange freqtrade configuration check exchange info exchange binance is officially supported by the freqtrade development team freqtrade configuration configuration info using pairlist from configuration freqtrade configuration config validation info validating configuration freqtrade commands optimize commands info starting freqtrade in backtesting mode freqtrade exchange exchange info instance is running with dry run enabled freqtrade exchange exchange info using ccxt freqtrade exchange exchange info using exchange binance freqtrade resolvers exchange resolver info using resolved exchange binance freqtrade resolvers iresolver info using resolved strategy broken result from freqtrade user data strategies broken result py freqtrade strategy hyper info loading parameters from file freqtrade user data strategies broken result json freqtrade resolvers strategy resolver info override strategy timeframe with value in config file freqtrade resolvers strategy resolver info override strategy stake currency with value in config file usdt freqtrade resolvers strategy resolver info override strategy stake amount with value in config file freqtrade resolvers strategy resolver info override strategy unfilledtimeout with value in config file entry exit exit timeout count unit minutes freqtrade resolvers strategy resolver info strategy using minimal roi freqtrade resolvers strategy resolver info strategy using timeframe freqtrade resolvers strategy resolver info strategy using stoploss freqtrade resolvers strategy resolver info strategy using trailing stop true freqtrade resolvers strategy resolver info strategy using trailing stop positive freqtrade resolvers strategy resolver info strategy using trailing stop positive offset freqtrade resolvers strategy resolver info strategy using trailing only offset is reached true freqtrade resolvers strategy resolver info strategy using use custom stoploss false freqtrade resolvers strategy resolver info strategy using process only new candles true freqtrade resolvers strategy resolver info strategy using order types entry limit exit limit stoploss limit stoploss on exchange false stoploss on exchange interval freqtrade resolvers strategy resolver info strategy using order time in force entry gtc exit gtc freqtrade resolvers strategy resolver info strategy using stake currency usdt freqtrade resolvers strategy resolver info strategy using stake amount freqtrade resolvers strategy resolver info strategy using protections freqtrade resolvers strategy resolver info strategy using startup candle count freqtrade resolvers strategy resolver info strategy using unfilledtimeout entry exit exit timeout count unit minutes freqtrade resolvers strategy resolver info strategy using use exit signal false freqtrade resolvers strategy resolver info strategy using exit profit only false freqtrade resolvers strategy resolver info strategy using ignore roi if entry signal false freqtrade resolvers strategy resolver info strategy using exit profit offset freqtrade resolvers strategy resolver info strategy using disable dataframe checks false freqtrade resolvers strategy resolver info strategy using ignore buying expired candle after freqtrade resolvers strategy resolver info strategy using position adjustment enable false freqtrade resolvers strategy resolver info strategy using max entry position adjustment freqtrade configuration config validation info validating configuration freqtrade resolvers iresolver info using resolved pairlist staticpairlist from freqtrade freqtrade plugins pairlist staticpairlist py freqtrade data history idatahandler warning shib usdt spot data starts at freqtrade data history idatahandler warning flow usdt spot data starts at freqtrade data history idatahandler warning icp usdt spot data starts at freqtrade data history idatahandler warning qnt usdt spot data starts at freqtrade optimize backtesting info loading data from up to days freqtrade optimize backtesting info dataload complete calculating indicators freqtrade optimize backtesting info running backtesting for strategy broken result freqtrade strategy hyper info strategy parameter bb time freqtrade strategy hyper info strategy parameter roc freqtrade strategy hyper info no params for sell found using default values freqtrade strategy hyper info no params for protection found using default values freqtrade optimize backtesting info backtesting with data from up to days freqtrade misc info dumping json to freqtrade user data backtest results backtest result meta json freqtrade misc info dumping json to freqtrade user data backtest results backtest result json freqtrade misc info dumping json to freqtrade user data backtest results last result json result for strategy broken result backtesting report pair buys avg profit cum profit tot profit usdt tot profit avg duration win draw loss win uni usdt fil usdt hnt usdt mana usdt axs usdt day ltc usdt day sand usdt eth usdt hbar usdt bch usdt day xrp usdt day avax usdt qnt usdt eos usdt icp usdt day xtz usdt day btc usdt day theta usdt dot usdt link usdt day shib usdt egld usdt day algo usdt xmr usdt day etc usdt day flow usdt days ftt usdt day vet usdt day atom usdt near usdt trx usdt day sol usdt day aave usdt matic usdt xlm usdt doge usdt ada usdt days total day enter tag stats tag buys avg profit cum profit tot profit usdt tot profit avg duration win draw loss win total day exit reason stats exit reason exits win draws loss win avg profit cum profit tot profit usdt tot profit roi stop loss trailing stop loss left open trades report pair buys avg profit cum profit tot profit usdt tot profit avg duration win draw loss win total summary metrics metric value backtesting from backtesting to max open trades total daily avg trades starting balance usdt final balance usdt absolute profit usdt total profit cagr profit factor trades per day avg daily profit avg stake amount usdt total trade volume usdt best pair uni usdt worst pair ada usdt best trade uni usdt worst trade ada usdt best day usdt worst day usdt days win draw lose avg duration winners day avg duration loser day rejected entry signals entry exit timeouts min balance usdt max balance usdt max of account underwater absolute drawdown account absolute drawdown usdt drawdown high usdt drawdown low usdt drawdown start drawdown end market change freqtrade exchange exchange info closing async ccxt session there s no changes between hyperopt and backtest same config file but results are different o example of the strategy where this happens is below i know it s bad but that s just an example do not remove these libs import numpy as np noqa import pandas as pd noqa import datetime from pandas import dataframe from freqtrade strategy import booleanparameter categoricalparameter decimalparameter istrategy intparameter merge informative pair add your lib to import here import talib import talib abstract as ta import pandas ta as pta import freqtrade vendor qtpylib indicators as qtpylib from technical util import resample to interval resampled merge class broken result istrategy roi table minimal roi stoploss stoploss trailing stop trailing stop true trailing stop positive trailing stop positive offset trailing only offset is reached true exit profit only true use exit signal false optimal timeframe for the strategy timeframe roc decimalparameter low high decimals default space buy optimize true bb time intparameter low high default space buy optimize true def populate indicators self dataframe dataframe metadata dict dataframe stochastic slow stoch ta stoch dataframe dataframe stoch dataframe stoch dataframe ta roc dataframe timeperiod bollinger qtpylib bollinger bands dataframe window self bb time value stds dataframe bollinger dataframe bollinger dataframe bollinger dataframe dataframe rolling self bb time value mean rolling std dataframe rolling self bb time value std dataframe dataframe rolling std dataframe dataframe rolling std dataframe ta max dataframe timeperiod return dataframe def populate buy trend self dataframe dataframe metadata dict dataframe dataframe loc dataframe dataframe shift dataframe self roc value qtpylib crossed above dataframe dataframe dataframe dataframe shift dataframe dataframe buy return dataframe def populate sell trend self dataframe dataframe metadata dict dataframe based on ta indicators populates the sell signal for the given dataframe param dataframe dataframe return dataframe with buy column dataframe loc sell return dataframe steps to reproduce run hyperopt for a few epochs and stop it or let it finish run backtest using generated json file or copying hyperopt parameters into the strategy observe observed results what happened results are different what did you expect to happen they should be the same notes i think hyperopt has a problem with optimizing the timeperiod
| 0
|
64,025
| 12,412,675,452
|
IssuesEvent
|
2020-05-22 11:04:01
|
topcoder-platform/community-app
|
https://api.github.com/repos/topcoder-platform/community-app
|
opened
|
Challenge Listings: Challenges not loading for subcommunities
|
Challenge Listing Test Env v5-intgration-sub-code
|
Challenge Listings: Challenges not loading for subcommunities
Example
https://test-community-app.topcoder-dev.com/__community__/blockchain/challenges?communityId=blockchain
<img width="1440" alt="Screenshot 2020-05-22 at 4 32 02 PM" src="https://user-images.githubusercontent.com/58783823/82661593-f37fb500-9c49-11ea-9c45-8a1c596a38b9.png">
<img width="1440" alt="Screenshot 2020-05-22 at 4 32 06 PM" src="https://user-images.githubusercontent.com/58783823/82661611-f9759600-9c49-11ea-85da-c68b6aedc049.png">
|
1.0
|
Challenge Listings: Challenges not loading for subcommunities - Challenge Listings: Challenges not loading for subcommunities
Example
https://test-community-app.topcoder-dev.com/__community__/blockchain/challenges?communityId=blockchain
<img width="1440" alt="Screenshot 2020-05-22 at 4 32 02 PM" src="https://user-images.githubusercontent.com/58783823/82661593-f37fb500-9c49-11ea-9c45-8a1c596a38b9.png">
<img width="1440" alt="Screenshot 2020-05-22 at 4 32 06 PM" src="https://user-images.githubusercontent.com/58783823/82661611-f9759600-9c49-11ea-85da-c68b6aedc049.png">
|
non_build
|
challenge listings challenges not loading for subcommunities challenge listings challenges not loading for subcommunities example img width alt screenshot at pm src img width alt screenshot at pm src
| 0
|
448,096
| 31,766,962,553
|
IssuesEvent
|
2023-09-12 09:21:26
|
appwrite/docs
|
https://api.github.com/repos/appwrite/docs
|
closed
|
📚 Documentation: GraphQL docs show REST endpoints
|
documentation
|
### 💭 Description
in the docs for databases / server when switching to GraphQL the REST endpoints are given, see https://appwrite.io/docs/server/databases?sdk=graphql-default
<img width="402" alt="image" src="https://user-images.githubusercontent.com/1942953/211330919-7cd83482-631b-43ab-adfc-c5efe356b067.png">
### 👀 Have you spent some time to check if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Code of Conduct?
- [X] I have read the [Code of Conduct](https://github.com/appwrite/appwrite/blob/HEAD/CODE_OF_CONDUCT.md)
|
1.0
|
📚 Documentation: GraphQL docs show REST endpoints - ### 💭 Description
in the docs for databases / server when switching to GraphQL the REST endpoints are given, see https://appwrite.io/docs/server/databases?sdk=graphql-default
<img width="402" alt="image" src="https://user-images.githubusercontent.com/1942953/211330919-7cd83482-631b-43ab-adfc-c5efe356b067.png">
### 👀 Have you spent some time to check if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Code of Conduct?
- [X] I have read the [Code of Conduct](https://github.com/appwrite/appwrite/blob/HEAD/CODE_OF_CONDUCT.md)
|
non_build
|
📚 documentation graphql docs show rest endpoints 💭 description in the docs for databases server when switching to graphql the rest endpoints are given see img width alt image src 👀 have you spent some time to check if this issue has been raised before i checked and didn t find similar issue 🏢 have you read the code of conduct i have read the
| 0
|
96,445
| 27,855,223,286
|
IssuesEvent
|
2023-03-20 22:10:13
|
envoyproxy/envoy
|
https://api.github.com/repos/envoyproxy/envoy
|
closed
|
N3w3r release available `com_github_wasmtime`: v6.0.0 (current: v2.0.2)
|
area/build no stalebot dependencies
|
Package Name: com_github_wasmtime@2.0.2
Current Version: v2.0.2@2022-11-10
Available Version: v6.0.0@2023-02-20
Upstream releases: https://github.com/bytecodealliance/wasmtime/releases
|
1.0
|
N3w3r release available `com_github_wasmtime`: v6.0.0 (current: v2.0.2) -
Package Name: com_github_wasmtime@2.0.2
Current Version: v2.0.2@2022-11-10
Available Version: v6.0.0@2023-02-20
Upstream releases: https://github.com/bytecodealliance/wasmtime/releases
|
build
|
release available com github wasmtime current package name com github wasmtime current version available version upstream releases
| 1
|
84,305
| 24,270,578,508
|
IssuesEvent
|
2022-09-28 09:56:20
|
Decathlon/vitamin-design
|
https://api.github.com/repos/Decathlon/vitamin-design
|
opened
|
[component] Accordion
|
documentation 📝 enhancement 🚀 web 🔵 android 🟢 ios 🟡 build 🏗
|
### Duplicates ❌
- [X] I have searched the existing issues
### Which Figma library is concerned?
Not related to one Figma library
### Summary 💡
Structure are used to define architecture of your interface and ensure consistency between pages and products
### Dependencies 📦
_No response_
### Examples 🌈
_No response_
### Motivation 🔦
_No response_
### New 🆕
- [ ] Validated in a grooming session
### Backlog 📋
- [ ] Has a good summary
- [ ] Has example link(s)
### Design in progress 🏗
- [ ] Analysis _(functional & technical)_
- [ ] Design refinement
- [ ] Dev qualification _(properties, breakpoints, tokens etc.)_
- [ ] Documentation refinement _(inside the Figma branch)_
### Design review 👀
- [ ] Figma branch merged
- [ ] Documentation review _(and move the documentation into Vitamin Documentation Figma file)
### Ready to dev 👍
- [ ] Issues are created in all repositories affected _(vitamin-<web,android,ios,compose>)_
### Dev in progress 💻
- [ ] All issues are resolved _(to replace with issues links)_
|
1.0
|
[component] Accordion - ### Duplicates ❌
- [X] I have searched the existing issues
### Which Figma library is concerned?
Not related to one Figma library
### Summary 💡
Structure are used to define architecture of your interface and ensure consistency between pages and products
### Dependencies 📦
_No response_
### Examples 🌈
_No response_
### Motivation 🔦
_No response_
### New 🆕
- [ ] Validated in a grooming session
### Backlog 📋
- [ ] Has a good summary
- [ ] Has example link(s)
### Design in progress 🏗
- [ ] Analysis _(functional & technical)_
- [ ] Design refinement
- [ ] Dev qualification _(properties, breakpoints, tokens etc.)_
- [ ] Documentation refinement _(inside the Figma branch)_
### Design review 👀
- [ ] Figma branch merged
- [ ] Documentation review _(and move the documentation into Vitamin Documentation Figma file)
### Ready to dev 👍
- [ ] Issues are created in all repositories affected _(vitamin-<web,android,ios,compose>)_
### Dev in progress 💻
- [ ] All issues are resolved _(to replace with issues links)_
|
build
|
accordion duplicates ❌ i have searched the existing issues which figma library is concerned not related to one figma library summary 💡 structure are used to define architecture of your interface and ensure consistency between pages and products dependencies 📦 no response examples 🌈 no response motivation 🔦 no response new 🆕 validated in a grooming session backlog 📋 has a good summary has example link s design in progress 🏗 analysis functional technical design refinement dev qualification properties breakpoints tokens etc documentation refinement inside the figma branch design review 👀 figma branch merged documentation review and move the documentation into vitamin documentation figma file ready to dev 👍 issues are created in all repositories affected vitamin dev in progress 💻 all issues are resolved to replace with issues links
| 1
|
125,295
| 10,339,680,539
|
IssuesEvent
|
2019-09-03 19:57:09
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: Firefox XPack UI Functional Tests.x-pack/test/functional/apps/infra/logs_source_configuration·ts - InfraOps app Logs Page with logs present renders the changed log columns with their headers
|
Team:infra-logs-ui [zube]: Ready failed-test
|
A test failed on a tracked branch
```
{ Error: expected [ 'Timestamp', 'host.name', '' ] to sort of equal [ 'host.name', 'Timestamp', '' ]
at Assertion.assert (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-firefoxSmoke/node/immutable/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.eql (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-firefoxSmoke/node/immutable/kibana/packages/kbn-expect/expect.js:235:8)
at Context.eql (test/functional/apps/infra/logs_source_configuration.ts:110:39)
at process._tickCallback (internal/process/next_tick.js:68:7)
actual: '[\n "Timestamp"\n "host.name"\n ""\n]',
expected: '[\n "host.name"\n "Timestamp"\n ""\n]',
showDiff: true }
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=x-pack-firefoxSmoke,node=immutable/1106/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Firefox XPack UI Functional Tests.x-pack/test/functional/apps/infra/logs_source_configuration·ts","test.name":"InfraOps app Logs Page with logs present renders the changed log columns with their headers","test.failCount":2}} -->
|
1.0
|
Failing test: Firefox XPack UI Functional Tests.x-pack/test/functional/apps/infra/logs_source_configuration·ts - InfraOps app Logs Page with logs present renders the changed log columns with their headers - A test failed on a tracked branch
```
{ Error: expected [ 'Timestamp', 'host.name', '' ] to sort of equal [ 'host.name', 'Timestamp', '' ]
at Assertion.assert (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-firefoxSmoke/node/immutable/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.eql (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-firefoxSmoke/node/immutable/kibana/packages/kbn-expect/expect.js:235:8)
at Context.eql (test/functional/apps/infra/logs_source_configuration.ts:110:39)
at process._tickCallback (internal/process/next_tick.js:68:7)
actual: '[\n "Timestamp"\n "host.name"\n ""\n]',
expected: '[\n "host.name"\n "Timestamp"\n ""\n]',
showDiff: true }
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=x-pack-firefoxSmoke,node=immutable/1106/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Firefox XPack UI Functional Tests.x-pack/test/functional/apps/infra/logs_source_configuration·ts","test.name":"InfraOps app Logs Page with logs present renders the changed log columns with their headers","test.failCount":2}} -->
|
non_build
|
failing test firefox xpack ui functional tests x pack test functional apps infra logs source configuration·ts infraops app logs page with logs present renders the changed log columns with their headers a test failed on a tracked branch error expected to sort of equal at assertion assert var lib jenkins workspace elastic kibana master job x pack firefoxsmoke node immutable kibana packages kbn expect expect js at assertion eql var lib jenkins workspace elastic kibana master job x pack firefoxsmoke node immutable kibana packages kbn expect expect js at context eql test functional apps infra logs source configuration ts at process tickcallback internal process next tick js actual expected showdiff true first failure
| 0
|
13,542
| 5,396,069,704
|
IssuesEvent
|
2017-02-27 10:33:56
|
IMA-WorldHealth/bhima-2.X
|
https://api.github.com/repos/IMA-WorldHealth/bhima-2.X
|
opened
|
(build) Set up automatic deployment to either local or remote servers
|
build process
|
As we spin up more and more production instances, we will need to ensure that we are able to successfully upgrade each production instance before investing lots of money traveling into the field.
I propose that we either spin up a local server that will automatically fetch and build the latest commit from master when merges happen. This instance should contain the latest database from each installation point to ensure that we are testing with real data. We can then build our migration scripts locally before going to the field and discover any bugs here.
|
1.0
|
(build) Set up automatic deployment to either local or remote servers - As we spin up more and more production instances, we will need to ensure that we are able to successfully upgrade each production instance before investing lots of money traveling into the field.
I propose that we either spin up a local server that will automatically fetch and build the latest commit from master when merges happen. This instance should contain the latest database from each installation point to ensure that we are testing with real data. We can then build our migration scripts locally before going to the field and discover any bugs here.
|
build
|
build set up automatic deployment to either local or remote servers as we spin up more and more production instances we will need to ensure that we are able to successfully upgrade each production instance before investing lots of money traveling into the field i propose that we either spin up a local server that will automatically fetch and build the latest commit from master when merges happen this instance should contain the latest database from each installation point to ensure that we are testing with real data we can then build our migration scripts locally before going to the field and discover any bugs here
| 1
|
33,836
| 4,863,100,079
|
IssuesEvent
|
2016-11-14 14:35:14
|
easydigitaldownloads/easy-digital-downloads
|
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
|
closed
|
TypeError: a(…).ajaxForm is not a function
|
Bug Has PR Needs Testing Priority: Low
|
> Upgraded to latest WP core (WordPress 4.6.1) and now when editing pages, this JS error is thrown, preventing other popups and actions on the page:
>
> “TypeError: a(…).ajaxForm is not a function”
> /wp-content/plugins/easy-digital-downloads/assests/js/admin-scripts.min.js:1:24505
>
> I am running latest EDD 2.6.8 with Stripe Payment Gateway add-on 2.5.4.
>
> Disabling EDD allows the correct execution of editing pages. Please advise.
**Source:** https://wordpress.org/support/topic/wordpress-4-6-1-js-error/
It look likes [`admin-scripts.js`](https://github.com/easydigitaldownloads/easy-digital-downloads/blob/2.6.11/assets/js/admin-scripts.js#L1561-L1567) requires [`jquery-form`](https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/js/jquery/jquery.form.js) for the jQuery [`ajaxForm`](https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/js/jquery/jquery.form.js#L789) function.
* https://github.com/easydigitaldownloads/easy-digital-downloads/blob/2.6.11/includes/scripts.php#L183-L195
* https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/js/jquery/jquery.form.js
* https://developer.wordpress.org/reference/functions/wp_enqueue_script/
* https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/script-loader.php#L236-L237
**Pull request:** https://github.com/easydigitaldownloads/easy-digital-downloads/pull/5205
|
1.0
|
TypeError: a(…).ajaxForm is not a function - > Upgraded to latest WP core (WordPress 4.6.1) and now when editing pages, this JS error is thrown, preventing other popups and actions on the page:
>
> “TypeError: a(…).ajaxForm is not a function”
> /wp-content/plugins/easy-digital-downloads/assests/js/admin-scripts.min.js:1:24505
>
> I am running latest EDD 2.6.8 with Stripe Payment Gateway add-on 2.5.4.
>
> Disabling EDD allows the correct execution of editing pages. Please advise.
**Source:** https://wordpress.org/support/topic/wordpress-4-6-1-js-error/
It look likes [`admin-scripts.js`](https://github.com/easydigitaldownloads/easy-digital-downloads/blob/2.6.11/assets/js/admin-scripts.js#L1561-L1567) requires [`jquery-form`](https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/js/jquery/jquery.form.js) for the jQuery [`ajaxForm`](https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/js/jquery/jquery.form.js#L789) function.
* https://github.com/easydigitaldownloads/easy-digital-downloads/blob/2.6.11/includes/scripts.php#L183-L195
* https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/js/jquery/jquery.form.js
* https://developer.wordpress.org/reference/functions/wp_enqueue_script/
* https://github.com/WordPress/WordPress/blob/4.6.1/wp-includes/script-loader.php#L236-L237
**Pull request:** https://github.com/easydigitaldownloads/easy-digital-downloads/pull/5205
|
non_build
|
typeerror a … ajaxform is not a function upgraded to latest wp core wordpress and now when editing pages this js error is thrown preventing other popups and actions on the page “typeerror a … ajaxform is not a function” wp content plugins easy digital downloads assests js admin scripts min js i am running latest edd with stripe payment gateway add on disabling edd allows the correct execution of editing pages please advise source it look likes requires for the jquery function pull request
| 0
|
13,796
| 5,451,991,228
|
IssuesEvent
|
2017-03-08 01:10:09
|
commercialhaskell/stack
|
https://api.github.com/repos/commercialhaskell/stack
|
closed
|
Stack ignores setup-depends altogether
|
component: build semantics in progress resolution: duplicate type: bug type: discuss
|
# Actual behavior
in [this log](http://lpaste.net/3330066059314069504), stack tries building `cairo` before `gtk2hs-buildtools`, and this fails. Luckily it builds `gtk2hs-buildtools`. Restarting the build then works.
# Expected behavior
stack should build `cairo` after `gtk2hs-buildtools`.
The dependency is declared by the cabal file but uses `setup-depends` (at least on HEAD)—this might trigger the bug:
https://github.com/gtk2hs/gtk2hs/blob/3fdb125c57e853a9ea473a54e277600be7d962bd/cairo/cairo.cabal#L46-L49
EDIT: this is split out from #2512.
|
1.0
|
Stack ignores setup-depends altogether - # Actual behavior
in [this log](http://lpaste.net/3330066059314069504), stack tries building `cairo` before `gtk2hs-buildtools`, and this fails. Luckily it builds `gtk2hs-buildtools`. Restarting the build then works.
# Expected behavior
stack should build `cairo` after `gtk2hs-buildtools`.
The dependency is declared by the cabal file but uses `setup-depends` (at least on HEAD)—this might trigger the bug:
https://github.com/gtk2hs/gtk2hs/blob/3fdb125c57e853a9ea473a54e277600be7d962bd/cairo/cairo.cabal#L46-L49
EDIT: this is split out from #2512.
|
build
|
stack ignores setup depends altogether actual behavior in stack tries building cairo before buildtools and this fails luckily it builds buildtools restarting the build then works expected behavior stack should build cairo after buildtools the dependency is declared by the cabal file but uses setup depends at least on head —this might trigger the bug edit this is split out from
| 1
|
32,442
| 6,793,503,326
|
IssuesEvent
|
2017-11-01 07:53:09
|
line/armeria
|
https://api.github.com/repos/line/armeria
|
closed
|
java.lang.ArrayIndexOutOfBoundsException: index out of covered value range
|
defect
|
Probably not a flaky test:
```
RollingHdrQuantilesTest.testThatConcurrentThreadsNotHungWithThreeChunks: FAILURE
com.linecorp.armeria.common.metric.RollingHdrQuantilesTest > testThatConcurrentThreadsNotHungWithThreeChunks FAILED
java.lang.ArrayIndexOutOfBoundsException: index out of covered value range
at org.HdrHistogram.AbstractHistogram.normalizeIndex(AbstractHistogram.java:2281)
at org.HdrHistogram.ConcurrentHistogram.recordConvertedDoubleValueWithCount(ConcurrentHistogram.java:185)
at org.HdrHistogram.DoubleHistogram.recordCountAtValue(DoubleHistogram.java:340)
at org.HdrHistogram.DoubleHistogram.recordValueWithCountAndExpectedInterval(DoubleHistogram.java:356)
at org.HdrHistogram.DoubleHistogram.recordValueWithExpectedInterval(DoubleHistogram.java:330)
at org.HdrHistogram.DoubleRecorder.recordValueWithExpectedInterval(DoubleRecorder.java:118)
at com.linecorp.armeria.common.metric.RollingHdrQuantiles.observe(RollingHdrQuantiles.java:114)
at com.linecorp.armeria.common.metric.RollingHdrQuantiles.observe(RollingHdrQuantiles.java:107)
at com.linecorp.armeria.common.metric.RollingHdrQuantilesTest.lambda$runInParallel$2(RollingHdrQuantilesTest.java:149)
```
|
1.0
|
java.lang.ArrayIndexOutOfBoundsException: index out of covered value range - Probably not a flaky test:
```
RollingHdrQuantilesTest.testThatConcurrentThreadsNotHungWithThreeChunks: FAILURE
com.linecorp.armeria.common.metric.RollingHdrQuantilesTest > testThatConcurrentThreadsNotHungWithThreeChunks FAILED
java.lang.ArrayIndexOutOfBoundsException: index out of covered value range
at org.HdrHistogram.AbstractHistogram.normalizeIndex(AbstractHistogram.java:2281)
at org.HdrHistogram.ConcurrentHistogram.recordConvertedDoubleValueWithCount(ConcurrentHistogram.java:185)
at org.HdrHistogram.DoubleHistogram.recordCountAtValue(DoubleHistogram.java:340)
at org.HdrHistogram.DoubleHistogram.recordValueWithCountAndExpectedInterval(DoubleHistogram.java:356)
at org.HdrHistogram.DoubleHistogram.recordValueWithExpectedInterval(DoubleHistogram.java:330)
at org.HdrHistogram.DoubleRecorder.recordValueWithExpectedInterval(DoubleRecorder.java:118)
at com.linecorp.armeria.common.metric.RollingHdrQuantiles.observe(RollingHdrQuantiles.java:114)
at com.linecorp.armeria.common.metric.RollingHdrQuantiles.observe(RollingHdrQuantiles.java:107)
at com.linecorp.armeria.common.metric.RollingHdrQuantilesTest.lambda$runInParallel$2(RollingHdrQuantilesTest.java:149)
```
|
non_build
|
java lang arrayindexoutofboundsexception index out of covered value range probably not a flaky test rollinghdrquantilestest testthatconcurrentthreadsnothungwiththreechunks failure com linecorp armeria common metric rollinghdrquantilestest testthatconcurrentthreadsnothungwiththreechunks failed java lang arrayindexoutofboundsexception index out of covered value range at org hdrhistogram abstracthistogram normalizeindex abstracthistogram java at org hdrhistogram concurrenthistogram recordconverteddoublevaluewithcount concurrenthistogram java at org hdrhistogram doublehistogram recordcountatvalue doublehistogram java at org hdrhistogram doublehistogram recordvaluewithcountandexpectedinterval doublehistogram java at org hdrhistogram doublehistogram recordvaluewithexpectedinterval doublehistogram java at org hdrhistogram doublerecorder recordvaluewithexpectedinterval doublerecorder java at com linecorp armeria common metric rollinghdrquantiles observe rollinghdrquantiles java at com linecorp armeria common metric rollinghdrquantiles observe rollinghdrquantiles java at com linecorp armeria common metric rollinghdrquantilestest lambda runinparallel rollinghdrquantilestest java
| 0
|
9,281
| 2,615,142,988
|
IssuesEvent
|
2015-03-01 06:18:13
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
Android problems
|
auto-migrated Priority-Medium Type-Defect
|
```
Please describe the issue:
On Android, you can move between the first few slides by swiping (good!) but
there's nothing there to tell you that (bad!), and when you get to the first
one with local database, you can no longer swipe to any new ones.
Please provide any additional information below. Nexus 1, Froyo FRF78
```
Original issue reported on code.google.com by `twb...@google.com` on 22 Jun 2010 at 11:08
|
1.0
|
Android problems - ```
Please describe the issue:
On Android, you can move between the first few slides by swiping (good!) but
there's nothing there to tell you that (bad!), and when you get to the first
one with local database, you can no longer swipe to any new ones.
Please provide any additional information below. Nexus 1, Froyo FRF78
```
Original issue reported on code.google.com by `twb...@google.com` on 22 Jun 2010 at 11:08
|
non_build
|
android problems please describe the issue on android you can move between the first few slides by swiping good but there s nothing there to tell you that bad and when you get to the first one with local database you can no longer swipe to any new ones please provide any additional information below nexus froyo original issue reported on code google com by twb google com on jun at
| 0
|
562,790
| 16,669,879,305
|
IssuesEvent
|
2021-06-07 09:31:04
|
epiphany-platform/epiphany
|
https://api.github.com/repos/epiphany-platform/epiphany
|
closed
|
[Backport][1.0.1] Filebeat communication error for AKS
|
priority/high type/backport
|
Backport task #2233 PR #2276
---
**DoD checklist**
* [x] Changelog updated
* [x] COMPONENTS.md updated / doesn't need to be updated
* [ ] Feature has automated tests
* [ ] Automated tests passed (QA pipelines)
* [ ] apply
* [ ] upgrade
* [ ] Idempotency tested
* [x] Documentation added / updated / doesn't need to be updated
* [x] All conversations in PR resolved
* [x] Solution meets requirements and is done according to design doc
* [x] Usage compliant with license
* [ ] Backport tasks created / doesn't need to be backported
|
1.0
|
[Backport][1.0.1] Filebeat communication error for AKS - Backport task #2233 PR #2276
---
**DoD checklist**
* [x] Changelog updated
* [x] COMPONENTS.md updated / doesn't need to be updated
* [ ] Feature has automated tests
* [ ] Automated tests passed (QA pipelines)
* [ ] apply
* [ ] upgrade
* [ ] Idempotency tested
* [x] Documentation added / updated / doesn't need to be updated
* [x] All conversations in PR resolved
* [x] Solution meets requirements and is done according to design doc
* [x] Usage compliant with license
* [ ] Backport tasks created / doesn't need to be backported
|
non_build
|
filebeat communication error for aks backport task pr dod checklist changelog updated components md updated doesn t need to be updated feature has automated tests automated tests passed qa pipelines apply upgrade idempotency tested documentation added updated doesn t need to be updated all conversations in pr resolved solution meets requirements and is done according to design doc usage compliant with license backport tasks created doesn t need to be backported
| 0
|
141,429
| 11,418,669,455
|
IssuesEvent
|
2020-02-03 05:31:50
|
bitcoin/bitcoin
|
https://api.github.com/repos/bitcoin/bitcoin
|
closed
|
Using Cirrus CI
|
Brainstorming Tests
|
**Is your feature request related to a problem? Please describe.**
The problem is that Travis does not do any stuff for BSD operating systems.
**Describe the solution you'd like**
Looks like [Cirrus CI](https://cirrus-ci.org/) supports (at least) FreeBSD.
Could this be added as a separate service as well?
**Additional context**
In general the Bitcoin Core support on the BSD operating systems is bad because no one really maintains it. This could go one step further to make Bitcoin Core available on onther platforms.
|
1.0
|
Using Cirrus CI - **Is your feature request related to a problem? Please describe.**
The problem is that Travis does not do any stuff for BSD operating systems.
**Describe the solution you'd like**
Looks like [Cirrus CI](https://cirrus-ci.org/) supports (at least) FreeBSD.
Could this be added as a separate service as well?
**Additional context**
In general the Bitcoin Core support on the BSD operating systems is bad because no one really maintains it. This could go one step further to make Bitcoin Core available on onther platforms.
|
non_build
|
using cirrus ci is your feature request related to a problem please describe the problem is that travis does not do any stuff for bsd operating systems describe the solution you d like looks like supports at least freebsd could this be added as a separate service as well additional context in general the bitcoin core support on the bsd operating systems is bad because no one really maintains it this could go one step further to make bitcoin core available on onther platforms
| 0
|
73,684
| 19,761,170,496
|
IssuesEvent
|
2022-01-16 12:47:06
|
chaotic-aur/packages
|
https://api.github.com/repos/chaotic-aur/packages
|
closed
|
[Outdated] python-manimpango
|
request:rebuild-pkg priority:high
|
### If available, link to the latest build
Build log site timed out multiple times.
### Package name
`python-manimpango`
### Latest build
`0.3.1-1`
### Latest version available
`0.4.0.post0-1`
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### More information
In my case, it's being used as a dependency for `emote`.
|
1.0
|
[Outdated] python-manimpango - ### If available, link to the latest build
Build log site timed out multiple times.
### Package name
`python-manimpango`
### Latest build
`0.3.1-1`
### Latest version available
`0.4.0.post0-1`
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### More information
In my case, it's being used as a dependency for `emote`.
|
build
|
python manimpango if available link to the latest build build log site timed out multiple times package name python manimpango latest build latest version available have you tested if the package builds in a clean chroot yes more information in my case it s being used as a dependency for emote
| 1
|
156,413
| 13,648,563,747
|
IssuesEvent
|
2020-09-26 09:54:36
|
AmazeeLabs/silverback-mono
|
https://api.github.com/repos/AmazeeLabs/silverback-mono
|
closed
|
FOUC when https://silverback.netlify.app/ docs site renders
|
bug documentation
|
*Background*
When https://silverback.netlify.app/ opens, there is a flash of unstyled content on page render.
*To replicate*
Simply browse to https://silverback.netlify.app/
*Expected behavior*
There is no unstyled content flash on page load
|
1.0
|
FOUC when https://silverback.netlify.app/ docs site renders - *Background*
When https://silverback.netlify.app/ opens, there is a flash of unstyled content on page render.
*To replicate*
Simply browse to https://silverback.netlify.app/
*Expected behavior*
There is no unstyled content flash on page load
|
non_build
|
fouc when docs site renders background when opens there is a flash of unstyled content on page render to replicate simply browse to expected behavior there is no unstyled content flash on page load
| 0
|
57,740
| 14,212,567,953
|
IssuesEvent
|
2020-11-17 00:23:18
|
chris-pearce/backpack.css
|
https://api.github.com/repos/chris-pearce/backpack.css
|
closed
|
Bring back the license comment in `index.css`
|
build ⚙️ next-major-version ✨ publishing 📤
|
```css
/*! backpack.css v2.2.0 | MIT License | github.com/chris-pearce/backpack.css */
```
https://github.com/sindresorhus/replace-in-files-cli#readme
|
1.0
|
Bring back the license comment in `index.css` - ```css
/*! backpack.css v2.2.0 | MIT License | github.com/chris-pearce/backpack.css */
```
https://github.com/sindresorhus/replace-in-files-cli#readme
|
build
|
bring back the license comment in index css css backpack css mit license github com chris pearce backpack css
| 1
|
276,020
| 30,320,029,130
|
IssuesEvent
|
2023-07-10 18:29:13
|
wrangleworks/WranglesPY
|
https://api.github.com/repos/wrangleworks/WranglesPY
|
opened
|
Allow parameterizing queries for recipe sql
|
Security
|
The current implementation allows writing raw SQL. If variables are included within the SQL this is open to SQL injection.
Pandas already supports defining parameters, so add this as an option.
https://pandas.pydata.org/docs/reference/api/pandas.read_sql.html
This is also used for the generic where method for wrangles/filter and write/connectors.
|
True
|
Allow parameterizing queries for recipe sql - The current implementation allows writing raw SQL. If variables are included within the SQL this is open to SQL injection.
Pandas already supports defining parameters, so add this as an option.
https://pandas.pydata.org/docs/reference/api/pandas.read_sql.html
This is also used for the generic where method for wrangles/filter and write/connectors.
|
non_build
|
allow parameterizing queries for recipe sql the current implementation allows writing raw sql if variables are included within the sql this is open to sql injection pandas already supports defining parameters so add this as an option this is also used for the generic where method for wrangles filter and write connectors
| 0
|
434,572
| 30,456,245,428
|
IssuesEvent
|
2023-07-16 23:06:34
|
UWB-Biocomputing/Graphitti
|
https://api.github.com/repos/UWB-Biocomputing/Graphitti
|
closed
|
Add override specifier
|
documentation refactor
|
C++ 11 introduced the `override` specifier which is used to explicitly indicate that a member function of a derived class is intended to override a virtual function from its base class.
There are multiple benefits of adding `override` specifier to our Graphitti codebase:
1. Improves code readability- The function that's meant to override from the base class is made very clear to the reader.
2. Enforces compiler checks- When a function is marked with `override`, the compiler will check whether the function is actually overriding a virtual function in the base class, providing an additional check.
3. Improve maintainability- When changes are made to the base class, `override` ensures that the derived classes are still correctly overriding the virtual functions.
Actions:
1. Add `override` specifier to all the derived classes where required.
2. Document the `override` feature in coding & cpp standards.
|
1.0
|
Add override specifier - C++ 11 introduced the `override` specifier which is used to explicitly indicate that a member function of a derived class is intended to override a virtual function from its base class.
There are multiple benefits of adding `override` specifier to our Graphitti codebase:
1. Improves code readability- The function that's meant to override from the base class is made very clear to the reader.
2. Enforces compiler checks- When a function is marked with `override`, the compiler will check whether the function is actually overriding a virtual function in the base class, providing an additional check.
3. Improve maintainability- When changes are made to the base class, `override` ensures that the derived classes are still correctly overriding the virtual functions.
Actions:
1. Add `override` specifier to all the derived classes where required.
2. Document the `override` feature in coding & cpp standards.
|
non_build
|
add override specifier c introduced the override specifier which is used to explicitly indicate that a member function of a derived class is intended to override a virtual function from its base class there are multiple benefits of adding override specifier to our graphitti codebase improves code readability the function that s meant to override from the base class is made very clear to the reader enforces compiler checks when a function is marked with override the compiler will check whether the function is actually overriding a virtual function in the base class providing an additional check improve maintainability when changes are made to the base class override ensures that the derived classes are still correctly overriding the virtual functions actions add override specifier to all the derived classes where required document the override feature in coding cpp standards
| 0
|
100,078
| 30,610,989,104
|
IssuesEvent
|
2023-07-23 15:56:36
|
kiwix/kiwix-js-windows
|
https://api.github.com/repos/kiwix/kiwix-js-windows
|
closed
|
Update the publish-appxbundle.yml workflow to include the new Electron-based appx
|
build
|
Now that the Microsoft Store is serving both the legacy appxbundle and the new Electron-based appx, the [publish-appxbundle.yml](https://github.com/kiwix/kiwix-js-windows/blob/main/.github/workflows/publish-appxbundle.yml) workflow needs to be updated to publish both apps to the [`kiwix-js-windows`](https://download.kiwix.org/release/kiwix-js-windows/) release directory.
|
1.0
|
Update the publish-appxbundle.yml workflow to include the new Electron-based appx - Now that the Microsoft Store is serving both the legacy appxbundle and the new Electron-based appx, the [publish-appxbundle.yml](https://github.com/kiwix/kiwix-js-windows/blob/main/.github/workflows/publish-appxbundle.yml) workflow needs to be updated to publish both apps to the [`kiwix-js-windows`](https://download.kiwix.org/release/kiwix-js-windows/) release directory.
|
build
|
update the publish appxbundle yml workflow to include the new electron based appx now that the microsoft store is serving both the legacy appxbundle and the new electron based appx the workflow needs to be updated to publish both apps to the release directory
| 1
|
117,914
| 25,215,394,314
|
IssuesEvent
|
2022-11-14 08:46:23
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
iOS: the AOT compiler crashes in .NET 8
|
untriaged area-Codegen-AOT-mono in-pr
|
### Description
The AOT compiler crashes during AOT compilation for some assemblies in .NET 8.
### Reproduction Steps
Test case:
[aottestcase-5de1b4f.zip](https://github.com/dotnet/runtime/files/9980646/aottestcase-5de1b4f.zip)
The test case contains a `repro.sh` script to reproduce the crash. You need to set the AOT_PACK environment variable to the location of your local AOT compiler pack for ios-arm64.
### Expected behavior
No crash
### Actual behavior
Crash:
```
$ AOT_PACK=/Users/rolf/work/maccore/net8.0/xamarin-macios/builds/downloads/dotnet-sdk-8.0.100-alpha.1.22553.12/packs/Microsoft.NETCore.App.Runtime.AOT.osx-x64.Cross.ios-arm64/8.0.0-alpha.1.22551.9 ./repro.sh
Mono Ahead of Time compiler - compiling assembly /Users/rolf/test/dotnet/aottestcase/dll/Touch.Client.dll
AOTID 4C1E041E-8395-B828-A40C-67B387B604D5
./repro.sh: line 26: 74901 Segmentation fault: 11 $AOT_COMPILER "--aot=mtriple=arm64-ios,data-outfile=$DATA_OUTFILE,static,asmonly,direct-icalls,full,readonly-value=ObjCRuntime.Runtime.Arch=i4/0,nodebug,dwarfdebug,llvm-path=$LLVM_PATH,outfile=$OUTFILE,llvm-outfile=$LLVM_OUTFILE" --debug --llvm -O=gsharedvt ./dll/"$assembly".dll
```
Crash report: https://gist.github.com/rolfbjarne/1380a700e6b9bb115dad33ae35369d98
### Regression?
Since .NET 7, yes
### Known Workarounds
_No response_
### Configuration
https://github.com/xamarin/xamarin-macios/blob/35fb9aca07fffdb14a6500be28f4f4d9984764ab/eng/Version.Details.xml#L3-L19
### Other information
_No response_
|
1.0
|
iOS: the AOT compiler crashes in .NET 8 - ### Description
The AOT compiler crashes during AOT compilation for some assemblies in .NET 8.
### Reproduction Steps
Test case:
[aottestcase-5de1b4f.zip](https://github.com/dotnet/runtime/files/9980646/aottestcase-5de1b4f.zip)
The test case contains a `repro.sh` script to reproduce the crash. You need to set the AOT_PACK environment variable to the location of your local AOT compiler pack for ios-arm64.
### Expected behavior
No crash
### Actual behavior
Crash:
```
$ AOT_PACK=/Users/rolf/work/maccore/net8.0/xamarin-macios/builds/downloads/dotnet-sdk-8.0.100-alpha.1.22553.12/packs/Microsoft.NETCore.App.Runtime.AOT.osx-x64.Cross.ios-arm64/8.0.0-alpha.1.22551.9 ./repro.sh
Mono Ahead of Time compiler - compiling assembly /Users/rolf/test/dotnet/aottestcase/dll/Touch.Client.dll
AOTID 4C1E041E-8395-B828-A40C-67B387B604D5
./repro.sh: line 26: 74901 Segmentation fault: 11 $AOT_COMPILER "--aot=mtriple=arm64-ios,data-outfile=$DATA_OUTFILE,static,asmonly,direct-icalls,full,readonly-value=ObjCRuntime.Runtime.Arch=i4/0,nodebug,dwarfdebug,llvm-path=$LLVM_PATH,outfile=$OUTFILE,llvm-outfile=$LLVM_OUTFILE" --debug --llvm -O=gsharedvt ./dll/"$assembly".dll
```
Crash report: https://gist.github.com/rolfbjarne/1380a700e6b9bb115dad33ae35369d98
### Regression?
Since .NET 7, yes
### Known Workarounds
_No response_
### Configuration
https://github.com/xamarin/xamarin-macios/blob/35fb9aca07fffdb14a6500be28f4f4d9984764ab/eng/Version.Details.xml#L3-L19
### Other information
_No response_
|
non_build
|
ios the aot compiler crashes in net description the aot compiler crashes during aot compilation for some assemblies in net reproduction steps test case the test case contains a repro sh script to reproduce the crash you need to set the aot pack environment variable to the location of your local aot compiler pack for ios expected behavior no crash actual behavior crash aot pack users rolf work maccore xamarin macios builds downloads dotnet sdk alpha packs microsoft netcore app runtime aot osx cross ios alpha repro sh mono ahead of time compiler compiling assembly users rolf test dotnet aottestcase dll touch client dll aotid repro sh line segmentation fault aot compiler aot mtriple ios data outfile data outfile static asmonly direct icalls full readonly value objcruntime runtime arch nodebug dwarfdebug llvm path llvm path outfile outfile llvm outfile llvm outfile debug llvm o gsharedvt dll assembly dll crash report regression since net yes known workarounds no response configuration other information no response
| 0
|
242,740
| 7,846,237,764
|
IssuesEvent
|
2018-06-19 14:58:58
|
google/google-api-dotnet-client
|
https://api.github.com/repos/google/google-api-dotnet-client
|
closed
|
.NET Framwork 4.7.2 Exception: System.Net.Http (works with 4.6.2)
|
priority: p2 status: investigating type: bug
|
> System.IO.FileLoadException: "Could not load file or assembly 'System.Net.Http, Version=4.2.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The located assembly's manifest definition does not match the assembly reference. (Exception from HRESULT: 0x80131040)"
It works with 4.6.2. I'm pretty certain about this issue
|
1.0
|
.NET Framwork 4.7.2 Exception: System.Net.Http (works with 4.6.2) - > System.IO.FileLoadException: "Could not load file or assembly 'System.Net.Http, Version=4.2.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The located assembly's manifest definition does not match the assembly reference. (Exception from HRESULT: 0x80131040)"
It works with 4.6.2. I'm pretty certain about this issue
|
non_build
|
net framwork exception system net http works with system io fileloadexception could not load file or assembly system net http version culture neutral publickeytoken or one of its dependencies the located assembly s manifest definition does not match the assembly reference exception from hresult it works with i m pretty certain about this issue
| 0
|
15,758
| 10,278,211,932
|
IssuesEvent
|
2019-08-25 12:35:08
|
ElektraInitiative/libelektra
|
https://api.github.com/repos/ElektraInitiative/libelektra
|
closed
|
kdb-gen: fail if no default supplied
|
bug help wanted usability
|
Applications are only guaranteed to get a key if they either require the key or provide an default.
The code generator should emit a warning if settings are neither, and do not emit a tag then. (Later we could provide some ways for `optional` keys, but we do not need it now.)
We should however test the specification with users.
|
True
|
kdb-gen: fail if no default supplied - Applications are only guaranteed to get a key if they either require the key or provide an default.
The code generator should emit a warning if settings are neither, and do not emit a tag then. (Later we could provide some ways for `optional` keys, but we do not need it now.)
We should however test the specification with users.
|
non_build
|
kdb gen fail if no default supplied applications are only guaranteed to get a key if they either require the key or provide an default the code generator should emit a warning if settings are neither and do not emit a tag then later we could provide some ways for optional keys but we do not need it now we should however test the specification with users
| 0
|
58,877
| 14,498,843,818
|
IssuesEvent
|
2020-12-11 16:00:03
|
eclipse-sirius/sirius-components
|
https://api.github.com/repos/eclipse-sirius/sirius-components
|
opened
|
Move to Spring Boot 2.4.1
|
area: backend area: build 🚧 difficulty: starter 👧 type: enhancement ✨
|
We've recently upgrade to Spring Boot 2.4.0, but 2.4.1 has just been released:
* https://spring.io/blog/2020/12/11/spring-boot-2-4-1-available-now
* https://github.com/spring-projects/spring-boot/releases/tag/v2.4.1
|
1.0
|
Move to Spring Boot 2.4.1 - We've recently upgrade to Spring Boot 2.4.0, but 2.4.1 has just been released:
* https://spring.io/blog/2020/12/11/spring-boot-2-4-1-available-now
* https://github.com/spring-projects/spring-boot/releases/tag/v2.4.1
|
build
|
move to spring boot we ve recently upgrade to spring boot but has just been released
| 1
|
81,870
| 15,817,862,030
|
IssuesEvent
|
2021-04-05 15:12:05
|
scdoja/suum
|
https://api.github.com/repos/scdoja/suum
|
closed
|
Component: Reusable Button Component
|
CODE: User-Interace
|
make a reusable button component that can be used throughout the app.

|
1.0
|
Component: Reusable Button Component - make a reusable button component that can be used throughout the app.

|
non_build
|
component reusable button component make a reusable button component that can be used throughout the app
| 0
|
286,201
| 21,566,572,026
|
IssuesEvent
|
2022-05-01 23:27:21
|
rancher/rke2
|
https://api.github.com/repos/rancher/rke2
|
opened
|
[Docs] Etcd backup/restore doesn't cover creating manual snapshots
|
kind/documentation
|
It appears the current doc page doesn't directly cover a manual snapshot example.
* https://docs.rke2.io/backup_restore/#creating-snapshots
For example on an RKE2 server node: `rke2 etcd-snapshot --name pre-upgrade-snapshot`
At the moment for anyone unfamiliar, it takes some digging via the command line help options to find this.
|
1.0
|
[Docs] Etcd backup/restore doesn't cover creating manual snapshots - It appears the current doc page doesn't directly cover a manual snapshot example.
* https://docs.rke2.io/backup_restore/#creating-snapshots
For example on an RKE2 server node: `rke2 etcd-snapshot --name pre-upgrade-snapshot`
At the moment for anyone unfamiliar, it takes some digging via the command line help options to find this.
|
non_build
|
etcd backup restore doesn t cover creating manual snapshots it appears the current doc page doesn t directly cover a manual snapshot example for example on an server node etcd snapshot name pre upgrade snapshot at the moment for anyone unfamiliar it takes some digging via the command line help options to find this
| 0
|
98,569
| 30,006,537,302
|
IssuesEvent
|
2023-06-26 12:49:51
|
Joe-Collins-1986/BitWise-DRF
|
https://api.github.com/repos/Joe-Collins-1986/BitWise-DRF
|
closed
|
USER STORY: LIST LINKS
|
EPIC - API Build MUST HAVE
|
As a **frontend developer** I can **access the links endpoint information for a specific article** so that **I can use the data to enhance the article page.**
**EPIC**
API Build
**Tasks:**
- [x] Model
- [x] View - Links list view for article
- [x] Serializer - Links serializer for article
- [x] URL - list
**Acceptance Criteria:**
- [x] API accessible via URL
- [x] Links relevant to article presented
|
1.0
|
USER STORY: LIST LINKS - As a **frontend developer** I can **access the links endpoint information for a specific article** so that **I can use the data to enhance the article page.**
**EPIC**
API Build
**Tasks:**
- [x] Model
- [x] View - Links list view for article
- [x] Serializer - Links serializer for article
- [x] URL - list
**Acceptance Criteria:**
- [x] API accessible via URL
- [x] Links relevant to article presented
|
build
|
user story list links as a frontend developer i can access the links endpoint information for a specific article so that i can use the data to enhance the article page epic api build tasks model view links list view for article serializer links serializer for article url list acceptance criteria api accessible via url links relevant to article presented
| 1
|
111,778
| 11,741,939,407
|
IssuesEvent
|
2020-03-11 23:05:48
|
ChilliCream/hotchocolate
|
https://api.github.com/repos/ChilliCream/hotchocolate
|
closed
|
.NET CLI Template produces error
|
documentation
|
**Describe the bug**
The project produced by the CLI does not run successfully.
**To Reproduce**
Steps to reproduce the behavior:
1. Following the documentation in https://hotchocolate.io/docs/dotnet-cli
2. Install template using "dotnet new -i HotChocolate.Templates.Server"
3. Generate a new project with "dotnet new graphql -n MyProjectName"
4. Run the project using Microsoft Visual Studio 2019 using IIS
5. Navigate to https://localhost:44373/playground/
6. Playground shows empty screen and query cannot be run

**Expected behavior**
https://localhost:44373/playground/ can be run and not show empty screen

**Desktop (please complete the following information):**
- Windows 10
|
1.0
|
.NET CLI Template produces error - **Describe the bug**
The project produced by the CLI does not run successfully.
**To Reproduce**
Steps to reproduce the behavior:
1. Following the documentation in https://hotchocolate.io/docs/dotnet-cli
2. Install template using "dotnet new -i HotChocolate.Templates.Server"
3. Generate a new project with "dotnet new graphql -n MyProjectName"
4. Run the project using Microsoft Visual Studio 2019 using IIS
5. Navigate to https://localhost:44373/playground/
6. Playground shows empty screen and query cannot be run

**Expected behavior**
https://localhost:44373/playground/ can be run and not show empty screen

**Desktop (please complete the following information):**
- Windows 10
|
non_build
|
net cli template produces error describe the bug the project produced by the cli does not run successfully to reproduce steps to reproduce the behavior following the documentation in install template using dotnet new i hotchocolate templates server generate a new project with dotnet new graphql n myprojectname run the project using microsoft visual studio using iis navigate to playground shows empty screen and query cannot be run expected behavior can be run and not show empty screen desktop please complete the following information windows
| 0
|
53,191
| 6,304,288,571
|
IssuesEvent
|
2017-07-21 15:34:25
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
JavaXCacheDependencyTest OOM failure
|
Team: Core Type: Test-Failure
|
```
05:14:10 Running classloading.JavaXCacheDependencyTest
05:15:05
05:15:05 Results :
05:15:05
05:15:05 Tests run: 14249, Failures: 0, Errors: 0, Skipped: 204
05:15:05
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [INFO] Reactor Summary:
05:15:05 [INFO]
05:15:05 [INFO] Hazelcast Root ..................................... SUCCESS [ 0.376 s]
05:15:05 [INFO] hazelcast .......................................... FAILURE [ 01:10 h]
05:15:05 [INFO] hazelcast-client ................................... SKIPPED
05:15:05 [INFO] hazelcast-spring ................................... SKIPPED
05:15:05 [INFO] hazelcast-build-utils .............................. SKIPPED
05:15:05 [INFO] hazelcast-all ...................................... SKIPPED
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [INFO] BUILD FAILURE
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [INFO] Total time: 01:10 h
05:15:05 [INFO] Finished at: 2017-03-26T05:15:05+03:00
05:15:05 [INFO] Final Memory: 70M/1691M
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test (default-test) on project hazelcast: Execution default-test of goal org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test failed: There was an error in the forked process
[ERROR] java.lang.OutOfMemoryError: Permian space
```
https://hazelcast-l337.ci.cloudbees.com/view/Hazelcast/job/Hazelcast-3.x-OracleJDK1.6-freestyle/55/consoleFull#-1360281990f83b0f40-1923-47eb-bce0-46f2cdd940e8
|
1.0
|
JavaXCacheDependencyTest OOM failure - ```
05:14:10 Running classloading.JavaXCacheDependencyTest
05:15:05
05:15:05 Results :
05:15:05
05:15:05 Tests run: 14249, Failures: 0, Errors: 0, Skipped: 204
05:15:05
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [INFO] Reactor Summary:
05:15:05 [INFO]
05:15:05 [INFO] Hazelcast Root ..................................... SUCCESS [ 0.376 s]
05:15:05 [INFO] hazelcast .......................................... FAILURE [ 01:10 h]
05:15:05 [INFO] hazelcast-client ................................... SKIPPED
05:15:05 [INFO] hazelcast-spring ................................... SKIPPED
05:15:05 [INFO] hazelcast-build-utils .............................. SKIPPED
05:15:05 [INFO] hazelcast-all ...................................... SKIPPED
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [INFO] BUILD FAILURE
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [INFO] Total time: 01:10 h
05:15:05 [INFO] Finished at: 2017-03-26T05:15:05+03:00
05:15:05 [INFO] Final Memory: 70M/1691M
05:15:05 [INFO] ------------------------------------------------------------------------
05:15:05 [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test (default-test) on project hazelcast: Execution default-test of goal org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test failed: There was an error in the forked process
[ERROR] java.lang.OutOfMemoryError: Permian space
```
https://hazelcast-l337.ci.cloudbees.com/view/Hazelcast/job/Hazelcast-3.x-OracleJDK1.6-freestyle/55/consoleFull#-1360281990f83b0f40-1923-47eb-bce0-46f2cdd940e8
|
non_build
|
javaxcachedependencytest oom failure running classloading javaxcachedependencytest results tests run failures errors skipped reactor summary hazelcast root success hazelcast failure hazelcast client skipped hazelcast spring skipped hazelcast build utils skipped hazelcast all skipped build failure total time h finished at final memory failed to execute goal org apache maven plugins maven surefire plugin test default test on project hazelcast execution default test of goal org apache maven plugins maven surefire plugin test failed there was an error in the forked process java lang outofmemoryerror permian space
| 0
|
10,767
| 2,964,217,032
|
IssuesEvent
|
2015-07-10 15:26:27
|
Growstuff/growstuff
|
https://api.github.com/repos/Growstuff/growstuff
|
closed
|
"Places" page should show plantings, harvests, seeds, etc
|
effort: large needs: design needs: frontend needs: visual design type: feature
|
It would be cool if http://growstuff.org/places/melbourne could show stuff other than members near here. For instance:
* what is growing/being harvested near here
* what seeds are available near here
In future we might even show seed retailers, and things like that.
I'd like to see a mockup/wireframe before implementation of this!
|
2.0
|
"Places" page should show plantings, harvests, seeds, etc - It would be cool if http://growstuff.org/places/melbourne could show stuff other than members near here. For instance:
* what is growing/being harvested near here
* what seeds are available near here
In future we might even show seed retailers, and things like that.
I'd like to see a mockup/wireframe before implementation of this!
|
non_build
|
places page should show plantings harvests seeds etc it would be cool if could show stuff other than members near here for instance what is growing being harvested near here what seeds are available near here in future we might even show seed retailers and things like that i d like to see a mockup wireframe before implementation of this
| 0
|
79,440
| 15,586,151,928
|
IssuesEvent
|
2021-03-18 01:17:25
|
jrshutske/unit-conversion-api
|
https://api.github.com/repos/jrshutske/unit-conversion-api
|
opened
|
CVE-2019-12086 (High) detected in jackson-databind-2.9.8.jar
|
security vulnerability
|
## CVE-2019-12086 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /unit-conversion-api/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library)
- spring-boot-actuator-autoconfigure-2.1.2.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x before 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint, the service has the mysql-connector-java jar (8.0.14 or earlier) in the classpath, and an attacker can host a crafted MySQL server reachable by the victim, an attacker can send a crafted JSON message that allows them to read arbitrary local files on the server. This occurs because of missing com.mysql.cj.jdbc.admin.MiniAdmin validation.
<p>Publish Date: 2019-05-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12086>CVE-2019-12086</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12086">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12086</a></p>
<p>Release Date: 2019-05-17</p>
<p>Fix Resolution: 2.9.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-12086 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2019-12086 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /unit-conversion-api/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library)
- spring-boot-actuator-autoconfigure-2.1.2.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x before 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint, the service has the mysql-connector-java jar (8.0.14 or earlier) in the classpath, and an attacker can host a crafted MySQL server reachable by the victim, an attacker can send a crafted JSON message that allows them to read arbitrary local files on the server. This occurs because of missing com.mysql.cj.jdbc.admin.MiniAdmin validation.
<p>Publish Date: 2019-05-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12086>CVE-2019-12086</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12086">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12086</a></p>
<p>Release Date: 2019-05-17</p>
<p>Fix Resolution: 2.9.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_build
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file unit conversion api pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter actuator release jar root library spring boot actuator autoconfigure release jar x jackson databind jar vulnerable library vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind x before when default typing is enabled either globally or for a specific property for an externally exposed json endpoint the service has the mysql connector java jar or earlier in the classpath and an attacker can host a crafted mysql server reachable by the victim an attacker can send a crafted json message that allows them to read arbitrary local files on the server this occurs because of missing com mysql cj jdbc admin miniadmin validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
40,284
| 10,489,996,997
|
IssuesEvent
|
2019-09-25 08:04:13
|
gradle/gradle
|
https://api.github.com/repos/gradle/gradle
|
closed
|
Show deprecation warnings for path sensitivity problems
|
@build-cache from:member
|
https://github.com/gradle/gradle/pull/10707 will start emitting deprecation warnings for most task validation problems detected during runtime. One problem sorely missing is to check if a task's input file properties have their normalization declared properly.
These validation warnings are only emitted for the static analysis right now.
---
cc: @gradle/build-cache
|
1.0
|
Show deprecation warnings for path sensitivity problems - https://github.com/gradle/gradle/pull/10707 will start emitting deprecation warnings for most task validation problems detected during runtime. One problem sorely missing is to check if a task's input file properties have their normalization declared properly.
These validation warnings are only emitted for the static analysis right now.
---
cc: @gradle/build-cache
|
build
|
show deprecation warnings for path sensitivity problems will start emitting deprecation warnings for most task validation problems detected during runtime one problem sorely missing is to check if a task s input file properties have their normalization declared properly these validation warnings are only emitted for the static analysis right now cc gradle build cache
| 1
|
12,859
| 15,107,922,083
|
IssuesEvent
|
2021-02-08 16:00:22
|
ckeditor/ckeditor5
|
https://api.github.com/repos/ckeditor/ckeditor5
|
closed
|
Crash when copying nested table which was pasted into the editor
|
domain:v4-compatibilty package:table squad:ux type:bug
|
## 📝 Provide detailed reproduction steps (if any)
1. Enable nested tables with the code:
```js
editor.model.schema.on( 'checkChild', ( evt, args ) => {
const context = args[ 0 ];
const childDefinition = args[ 1 ];
if ( context.endsWith( 'tableCell' ) && childDefinition && childDefinition.name == 'table' ) {
// Prevent next listeners from being called.
evt.stop();
// Set the checkChild()'s return value.
evt.return = true;
}
}, { priority: 'highest' } );
```
2. Create a table with 1 table inside
3. Copy and paste thes table somewhere in the editor
4. Select the pasted table and press `ctrl / cmd + c`
### ✔️ Expected result
The second table is copied, no errors.
### ❌ Actual result
Error in the console:
```js
Uncaught TypeError: Cannot read property 'getAttributeKeys' of null
at hasAnyAttribute (downcast.js:523)
at createViewTableCellElement (downcast.js:382)
at DowncastDispatcher.<anonymous> (downcast.js:69)
at DowncastDispatcher.fire (emittermixin.js:221)
at DowncastDispatcher._testAndFire (downcastdispatcher.js:560)
at DowncastDispatcher._convertInsertWithAttributes (downcastdispatcher.js:582)
at DowncastDispatcher.convertInsert (downcastdispatcher.js:197)
at DataController.toView (datacontroller.js:248)
at Document.onCopyCut (clipboard.js:151)
at Document.fire (emittermixin.js:221)
```
https://user-images.githubusercontent.com/34380544/105724525-3e17a680-5f28-11eb-8cbb-8e45a1594be5.mp4
## 📃 Other details
Related - https://github.com/ckeditor/ckeditor5/issues/3232.
---
If you'd like to see this fixed sooner, add a 👍 reaction to this post.
|
True
|
Crash when copying nested table which was pasted into the editor - ## 📝 Provide detailed reproduction steps (if any)
1. Enable nested tables with the code:
```js
editor.model.schema.on( 'checkChild', ( evt, args ) => {
const context = args[ 0 ];
const childDefinition = args[ 1 ];
if ( context.endsWith( 'tableCell' ) && childDefinition && childDefinition.name == 'table' ) {
// Prevent next listeners from being called.
evt.stop();
// Set the checkChild()'s return value.
evt.return = true;
}
}, { priority: 'highest' } );
```
2. Create a table with 1 table inside
3. Copy and paste thes table somewhere in the editor
4. Select the pasted table and press `ctrl / cmd + c`
### ✔️ Expected result
The second table is copied, no errors.
### ❌ Actual result
Error in the console:
```js
Uncaught TypeError: Cannot read property 'getAttributeKeys' of null
at hasAnyAttribute (downcast.js:523)
at createViewTableCellElement (downcast.js:382)
at DowncastDispatcher.<anonymous> (downcast.js:69)
at DowncastDispatcher.fire (emittermixin.js:221)
at DowncastDispatcher._testAndFire (downcastdispatcher.js:560)
at DowncastDispatcher._convertInsertWithAttributes (downcastdispatcher.js:582)
at DowncastDispatcher.convertInsert (downcastdispatcher.js:197)
at DataController.toView (datacontroller.js:248)
at Document.onCopyCut (clipboard.js:151)
at Document.fire (emittermixin.js:221)
```
https://user-images.githubusercontent.com/34380544/105724525-3e17a680-5f28-11eb-8cbb-8e45a1594be5.mp4
## 📃 Other details
Related - https://github.com/ckeditor/ckeditor5/issues/3232.
---
If you'd like to see this fixed sooner, add a 👍 reaction to this post.
|
non_build
|
crash when copying nested table which was pasted into the editor 📝 provide detailed reproduction steps if any enable nested tables with the code js editor model schema on checkchild evt args const context args const childdefinition args if context endswith tablecell childdefinition childdefinition name table prevent next listeners from being called evt stop set the checkchild s return value evt return true priority highest create a table with table inside copy and paste thes table somewhere in the editor select the pasted table and press ctrl cmd c ✔️ expected result the second table is copied no errors ❌ actual result error in the console js uncaught typeerror cannot read property getattributekeys of null at hasanyattribute downcast js at createviewtablecellelement downcast js at downcastdispatcher downcast js at downcastdispatcher fire emittermixin js at downcastdispatcher testandfire downcastdispatcher js at downcastdispatcher convertinsertwithattributes downcastdispatcher js at downcastdispatcher convertinsert downcastdispatcher js at datacontroller toview datacontroller js at document oncopycut clipboard js at document fire emittermixin js 📃 other details related if you d like to see this fixed sooner add a 👍 reaction to this post
| 0
|
81,441
| 23,463,947,230
|
IssuesEvent
|
2022-08-16 15:10:13
|
dotnet/arcade
|
https://api.github.com/repos/dotnet/arcade
|
closed
|
Build failed: dotnet-arcade-validation-official/main #20220816.1
|
Build Failed
|
Build [#20220816.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1946303) failed
## :x: : internal / dotnet-arcade-validation-official failed
### Summary
**Finished** - Tue, 16 Aug 2022 09:39:02 GMT
**Duration** - 37 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Validate Build Assets
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1946303/logs/299) - .packages\microsoft.dotnet.arcade.sdk\7.0.0-beta.22415.8\tools\SdkTasks\SigningValidation.proj(0,0): error : Version 7.0.100-preview.7.22377.5 of the .NET SDK requires at least version 17.2.0 of MSBuild. The current available version of MSBuild is 17.1.0.7609. Change the .NET SDK specified in global.json to an older version that requires the MSBuild version currently available.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1946303/logs/299) - .packages\microsoft.dotnet.arcade.sdk\7.0.0-beta.22415.8\tools\SdkTasks\SigningValidation.proj(0,0): error MSB4236: The SDK 'Microsoft.NET.Sdk' specified could not be found.
### Changes
- [11093bd2](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/11093bd2fe1bcad4c95631e0e36d729b60e876e6) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.8 (#3298)
- [06688fa8](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/06688fa88339f0e2463da07e7177c179b4633d3b) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.7 (#3297)
- [ae3b3ebe](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/ae3b3ebeffbdca3ed238b32f6df1035a44cfc8b8) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.5 (#3296)
- [f1cd51f8](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/f1cd51f878c81467a5c31f43fecab55022d07036) - Jon Fortescue - Switch to new images (#3294)
- [77e984e2](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/77e984e2fc74daeec6b6e24cafa7829d93fc18bb) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.3 (#3295)
- [b677103e](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/b677103ecfc2b12e096ed5ff860cb8a3b71b1c0a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.2 (#3292)
- [99ee9d69](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/99ee9d69cf08b869254b6c4c6ceecf7e7100298a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.1 (#3291)
- [8f86f1ae](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/8f86f1ae34238078e7c3fb56dbe3e20a3b6ddedc) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220812.5 (#3290)
- [fc279f47](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/fc279f47f882b8baedbe376e7ddc9854c7072aea) - dotnet-maestro[bot] - [main] Update dependencies from dotnet/arcade (#3287)
|
1.0
|
Build failed: dotnet-arcade-validation-official/main #20220816.1 - Build [#20220816.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1946303) failed
## :x: : internal / dotnet-arcade-validation-official failed
### Summary
**Finished** - Tue, 16 Aug 2022 09:39:02 GMT
**Duration** - 37 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Validate Build Assets
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1946303/logs/299) - .packages\microsoft.dotnet.arcade.sdk\7.0.0-beta.22415.8\tools\SdkTasks\SigningValidation.proj(0,0): error : Version 7.0.100-preview.7.22377.5 of the .NET SDK requires at least version 17.2.0 of MSBuild. The current available version of MSBuild is 17.1.0.7609. Change the .NET SDK specified in global.json to an older version that requires the MSBuild version currently available.
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1946303/logs/299) - .packages\microsoft.dotnet.arcade.sdk\7.0.0-beta.22415.8\tools\SdkTasks\SigningValidation.proj(0,0): error MSB4236: The SDK 'Microsoft.NET.Sdk' specified could not be found.
### Changes
- [11093bd2](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/11093bd2fe1bcad4c95631e0e36d729b60e876e6) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.8 (#3298)
- [06688fa8](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/06688fa88339f0e2463da07e7177c179b4633d3b) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.7 (#3297)
- [ae3b3ebe](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/ae3b3ebeffbdca3ed238b32f6df1035a44cfc8b8) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.5 (#3296)
- [f1cd51f8](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/f1cd51f878c81467a5c31f43fecab55022d07036) - Jon Fortescue - Switch to new images (#3294)
- [77e984e2](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/77e984e2fc74daeec6b6e24cafa7829d93fc18bb) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.3 (#3295)
- [b677103e](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/b677103ecfc2b12e096ed5ff860cb8a3b71b1c0a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.2 (#3292)
- [99ee9d69](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/99ee9d69cf08b869254b6c4c6ceecf7e7100298a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220815.1 (#3291)
- [8f86f1ae](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/8f86f1ae34238078e7c3fb56dbe3e20a3b6ddedc) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220812.5 (#3290)
- [fc279f47](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/fc279f47f882b8baedbe376e7ddc9854c7072aea) - dotnet-maestro[bot] - [main] Update dependencies from dotnet/arcade (#3287)
|
build
|
build failed dotnet arcade validation official main build failed x internal dotnet arcade validation official failed summary finished tue aug gmt duration minutes requested for dotnet bot reason batchedci details validate build assets x packages microsoft dotnet arcade sdk beta tools sdktasks signingvalidation proj error version preview of the net sdk requires at least version of msbuild the current available version of msbuild is change the net sdk specified in global json to an older version that requires the msbuild version currently available x packages microsoft dotnet arcade sdk beta tools sdktasks signingvalidation proj error the sdk microsoft net sdk specified could not be found changes dotnet maestro update dependencies from build dotnet maestro update dependencies from build dotnet maestro update dependencies from build jon fortescue switch to new images dotnet maestro update dependencies from build dotnet maestro update dependencies from build dotnet maestro update dependencies from build dotnet maestro update dependencies from build dotnet maestro update dependencies from dotnet arcade
| 1
|
46,878
| 11,903,634,598
|
IssuesEvent
|
2020-03-30 15:37:05
|
trilinos/Trilinos
|
https://api.github.com/repos/trilinos/Trilinos
|
opened
|
ATDM Trilinos configure broken on all CTS-1 and TLCC-2 platforms starting 2020-03-28
|
ATDM Sev: Critical client: ATDM impacting: build type: bug
|
## Next Action Status
???
## Description
As shown in [this query](https://testing-dev.sandia.gov/cdash/index.php?project=Trilinos&begin=2020-03-01&end=2020-03-30&filtercount=2&showfilters=1&filtercombine=or&field1=buildname&compare1=65&value1=Trilinos-atdm-tlcc2&field2=buildname&compare2=65&value2=Trilinos-atdm-cts1) all of the builds on 'cts1' and 'tlcc2' platforms have had configure failures starting 2028-03-28. The configure failures show:
```
-- The C compiler identification is Intel 19.0.5.20190815
-- Check for working C compiler: /opt/openmpi/4.0/intel/bin/mpicc
-- Check for working C compiler: /opt/openmpi/4.0/intel/bin/mpicc -- broken
CMake Error at /projects/cmake/toss3/3.12.2/share/cmake-3.12/Modules/CMakeTestCCompiler.cmake:52 (message):
The C compiler
"/opt/openmpi/4.0/intel/bin/mpicc"
is not able to compile a simple test program.
It fails with the following output:
Change Dir: /nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/SRC_AND_BUILD/BUILD/CMakeFiles/CMakeTmp
Run Build Command:"/projects/sems/install/toss3/sems/utility/ninja_fortran/1.8.2/bin/ninja" "cmTC_7b3d7"
[1/2] Building C object CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o
FAILED: CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o
/opt/openmpi/4.0/intel/bin/mpicc -MD -MT CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o -MF CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o.d -o CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o -c testCCompiler.c
/opt/openmpi/4.0/intel/bin/mpicc: error while loading shared libraries: libimf.so: cannot open shared object file: No such file or directory
ninja: build stopped: subcommand failed.
CMake will not be able to correctly generate this project.
Call Stack (most recent call first):
/nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/Trilinos/cmake/tribits/core/package_arch/TribitsGlobalMacros.cmake:1920 (ENABLE_LANGUAGE)
/nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/Trilinos/cmake/tribits/core/package_arch/TribitsProjectImpl.cmake:192 (TRIBITS_SETUP_ENV)
/nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/Trilinos/cmake/tribits/core/package_arch/TribitsProject.cmake:93 (TRIBITS_PROJECT_IMPL)
CMakeLists.txt:90 (TRIBITS_PROJECT)
```
|
1.0
|
ATDM Trilinos configure broken on all CTS-1 and TLCC-2 platforms starting 2020-03-28 - ## Next Action Status
???
## Description
As shown in [this query](https://testing-dev.sandia.gov/cdash/index.php?project=Trilinos&begin=2020-03-01&end=2020-03-30&filtercount=2&showfilters=1&filtercombine=or&field1=buildname&compare1=65&value1=Trilinos-atdm-tlcc2&field2=buildname&compare2=65&value2=Trilinos-atdm-cts1) all of the builds on 'cts1' and 'tlcc2' platforms have had configure failures starting 2028-03-28. The configure failures show:
```
-- The C compiler identification is Intel 19.0.5.20190815
-- Check for working C compiler: /opt/openmpi/4.0/intel/bin/mpicc
-- Check for working C compiler: /opt/openmpi/4.0/intel/bin/mpicc -- broken
CMake Error at /projects/cmake/toss3/3.12.2/share/cmake-3.12/Modules/CMakeTestCCompiler.cmake:52 (message):
The C compiler
"/opt/openmpi/4.0/intel/bin/mpicc"
is not able to compile a simple test program.
It fails with the following output:
Change Dir: /nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/SRC_AND_BUILD/BUILD/CMakeFiles/CMakeTmp
Run Build Command:"/projects/sems/install/toss3/sems/utility/ninja_fortran/1.8.2/bin/ninja" "cmTC_7b3d7"
[1/2] Building C object CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o
FAILED: CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o
/opt/openmpi/4.0/intel/bin/mpicc -MD -MT CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o -MF CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o.d -o CMakeFiles/cmTC_7b3d7.dir/testCCompiler.c.o -c testCCompiler.c
/opt/openmpi/4.0/intel/bin/mpicc: error while loading shared libraries: libimf.so: cannot open shared object file: No such file or directory
ninja: build stopped: subcommand failed.
CMake will not be able to correctly generate this project.
Call Stack (most recent call first):
/nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/Trilinos/cmake/tribits/core/package_arch/TribitsGlobalMacros.cmake:1920 (ENABLE_LANGUAGE)
/nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/Trilinos/cmake/tribits/core/package_arch/TribitsProjectImpl.cmake:192 (TRIBITS_SETUP_ENV)
/nscratch/jenkins/eclipse-slave/workspace/Trilinos-atdm-cts1-intel-19.0.5_openmpi-4.0.1_openmp_static_opt/Trilinos/cmake/tribits/core/package_arch/TribitsProject.cmake:93 (TRIBITS_PROJECT_IMPL)
CMakeLists.txt:90 (TRIBITS_PROJECT)
```
|
build
|
atdm trilinos configure broken on all cts and tlcc platforms starting next action status description as shown in all of the builds on and platforms have had configure failures starting the configure failures show the c compiler identification is intel check for working c compiler opt openmpi intel bin mpicc check for working c compiler opt openmpi intel bin mpicc broken cmake error at projects cmake share cmake modules cmaketestccompiler cmake message the c compiler opt openmpi intel bin mpicc is not able to compile a simple test program it fails with the following output change dir nscratch jenkins eclipse slave workspace trilinos atdm intel openmpi openmp static opt src and build build cmakefiles cmaketmp run build command projects sems install sems utility ninja fortran bin ninja cmtc building c object cmakefiles cmtc dir testccompiler c o failed cmakefiles cmtc dir testccompiler c o opt openmpi intel bin mpicc md mt cmakefiles cmtc dir testccompiler c o mf cmakefiles cmtc dir testccompiler c o d o cmakefiles cmtc dir testccompiler c o c testccompiler c opt openmpi intel bin mpicc error while loading shared libraries libimf so cannot open shared object file no such file or directory ninja build stopped subcommand failed cmake will not be able to correctly generate this project call stack most recent call first nscratch jenkins eclipse slave workspace trilinos atdm intel openmpi openmp static opt trilinos cmake tribits core package arch tribitsglobalmacros cmake enable language nscratch jenkins eclipse slave workspace trilinos atdm intel openmpi openmp static opt trilinos cmake tribits core package arch tribitsprojectimpl cmake tribits setup env nscratch jenkins eclipse slave workspace trilinos atdm intel openmpi openmp static opt trilinos cmake tribits core package arch tribitsproject cmake tribits project impl cmakelists txt tribits project
| 1
|
248,808
| 18,858,120,784
|
IssuesEvent
|
2021-11-12 09:24:29
|
James-Kua/pe
|
https://api.github.com/repos/James-Kua/pe
|
opened
|
Missing code blocks for keywords in command
|
type.DocumentationBug severity.Low
|
Should these be in code blocks, instead of double quotation marks?

<!--session: 1636703620420-0b3c217d-6d4e-49b3-84c2-0ff0684b2426-->
<!--Version: Web v3.4.1-->
|
1.0
|
Missing code blocks for keywords in command - Should these be in code blocks, instead of double quotation marks?

<!--session: 1636703620420-0b3c217d-6d4e-49b3-84c2-0ff0684b2426-->
<!--Version: Web v3.4.1-->
|
non_build
|
missing code blocks for keywords in command should these be in code blocks instead of double quotation marks
| 0
|
19,142
| 10,326,357,172
|
IssuesEvent
|
2019-09-02 01:32:55
|
yaeljacobs67/cncjs
|
https://api.github.com/repos/yaeljacobs67/cncjs
|
opened
|
CVE-2018-20677 (Medium) detected in bootstrap-3.3.7.tgz
|
security vulnerability
|
## CVE-2018-20677 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.tgz</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz</a></p>
<p>Path to dependency file: /cncjs/package.json</p>
<p>Path to vulnerable library: /cncjs/node_modules/bootstrap/package.json</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677>CVE-2018-20677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
</details>
<p></p>
|
True
|
CVE-2018-20677 (Medium) detected in bootstrap-3.3.7.tgz - ## CVE-2018-20677 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.tgz</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz</a></p>
<p>Path to dependency file: /cncjs/package.json</p>
<p>Path to vulnerable library: /cncjs/node_modules/bootstrap/package.json</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677>CVE-2018-20677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
</details>
<p></p>
|
non_build
|
cve medium detected in bootstrap tgz cve medium severity vulnerability vulnerable library bootstrap tgz the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file cncjs package json path to vulnerable library cncjs node modules bootstrap package json dependency hierarchy x bootstrap tgz vulnerable library vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter
| 0
|
19,682
| 10,416,745,770
|
IssuesEvent
|
2019-09-14 16:01:27
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Vulnerability roundup 73: pango-1.42.4: 1 advisory
|
1.severity: security
|
[search](https://search.nix.gsc.io/?q=pango&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=pango+in%3Apath&type=Code)
* [ ] [CVE-2019-1010238](https://nvd.nist.gov/vuln/detail/CVE-2019-1010238) (nixos-19.03)
Scanned versions: nixos-19.03: 31d476b8797. May contain false positives.
|
True
|
Vulnerability roundup 73: pango-1.42.4: 1 advisory - [search](https://search.nix.gsc.io/?q=pango&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=pango+in%3Apath&type=Code)
* [ ] [CVE-2019-1010238](https://nvd.nist.gov/vuln/detail/CVE-2019-1010238) (nixos-19.03)
Scanned versions: nixos-19.03: 31d476b8797. May contain false positives.
|
non_build
|
vulnerability roundup pango advisory nixos scanned versions nixos may contain false positives
| 0
|
11,788
| 5,088,533,373
|
IssuesEvent
|
2016-12-31 21:58:17
|
angular/angular-cli
|
https://api.github.com/repos/angular/angular-cli
|
closed
|
maintain the scripts order as per the angular-cli.json configuration
|
command: build need: investigation priority: 1 (urgent) type: bug
|
### OS?
Windows 10
### Versions.
angular-cli: 1.0.0-beta.24
node: 6.9.2
os: win32 x64
### Repro steps.
My `angular-cli.json` has the following:
```json
...
"scripts": [
...
"../node_modules/jquery-ui/ui/widget.js",
"../node_modules/jquery-ui/ui/widgets/datepicker.js",
"../node_modules/jquery-ui-timepicker-addon/dist/jquery-ui-timepicker-addon.js",
...
],
...
```
### The log given by the failure.
At https://github.com/trentrichardson/jQuery-Timepicker-Addon/blob/master/dist/jquery-ui-timepicker-addon.js#L15 I get Uncaught TypeError: Cannot read property 'timepicker' of undefined. I found this is because webpack is loading the scripts in the following order:
-jquery-ui-timepicker-addon.js
-widget.js
-datepicker.js
widget.js needs to load before jquery-ui-timepicker-addon.js. Why dose webpack not load in the same order as the configuration and how do you determine the order the files load?
### Mention any other details that might be useful.
http://stackoverflow.com/questions/39964281/how-to-order-angular-cli-scripts
|
1.0
|
maintain the scripts order as per the angular-cli.json configuration - ### OS?
Windows 10
### Versions.
angular-cli: 1.0.0-beta.24
node: 6.9.2
os: win32 x64
### Repro steps.
My `angular-cli.json` has the following:
```json
...
"scripts": [
...
"../node_modules/jquery-ui/ui/widget.js",
"../node_modules/jquery-ui/ui/widgets/datepicker.js",
"../node_modules/jquery-ui-timepicker-addon/dist/jquery-ui-timepicker-addon.js",
...
],
...
```
### The log given by the failure.
At https://github.com/trentrichardson/jQuery-Timepicker-Addon/blob/master/dist/jquery-ui-timepicker-addon.js#L15 I get Uncaught TypeError: Cannot read property 'timepicker' of undefined. I found this is because webpack is loading the scripts in the following order:
-jquery-ui-timepicker-addon.js
-widget.js
-datepicker.js
widget.js needs to load before jquery-ui-timepicker-addon.js. Why dose webpack not load in the same order as the configuration and how do you determine the order the files load?
### Mention any other details that might be useful.
http://stackoverflow.com/questions/39964281/how-to-order-angular-cli-scripts
|
build
|
maintain the scripts order as per the angular cli json configuration os windows versions angular cli beta node os repro steps my angular cli json has the following json scripts node modules jquery ui ui widget js node modules jquery ui ui widgets datepicker js node modules jquery ui timepicker addon dist jquery ui timepicker addon js the log given by the failure at i get uncaught typeerror cannot read property timepicker of undefined i found this is because webpack is loading the scripts in the following order jquery ui timepicker addon js widget js datepicker js widget js needs to load before jquery ui timepicker addon js why dose webpack not load in the same order as the configuration and how do you determine the order the files load mention any other details that might be useful
| 1
|
35,994
| 9,694,534,057
|
IssuesEvent
|
2019-05-24 19:14:17
|
psi4/psi4
|
https://api.github.com/repos/psi4/psi4
|
closed
|
Fails to find libint: it doesn't provide .cmake or .pc files
|
build
|
```external/upstream/libint/CMakeLists.txt``` has this line:
```
find_package(Libint 1.2.0 CONFIG QUIET COMPONENTS ${MAX_AM_ERI})
```
I have the latest version of libint installed: libint-1.2.1_1 (also see here https://repology.org/project/libint/versions).
I changed this line to the latest version:
```
find_package(Libint 1.2.1 CONFIG QUIET COMPONENTS ${MAX_AM_ERI})
```
but it fails to find it:
```
-- Suitable Libint could not be located, Building Libint instead.
```
libint-1.2.1_1 doesn't provide .pc or .cmake files, so I am not sure how can it even ever be able to find it through a cmake script like this?
It's undesirable to download and rebuild libint when the package is available.
Please also note that the latest generation of libint is libint2, which does provide ```libint2.pc``` .
|
1.0
|
Fails to find libint: it doesn't provide .cmake or .pc files - ```external/upstream/libint/CMakeLists.txt``` has this line:
```
find_package(Libint 1.2.0 CONFIG QUIET COMPONENTS ${MAX_AM_ERI})
```
I have the latest version of libint installed: libint-1.2.1_1 (also see here https://repology.org/project/libint/versions).
I changed this line to the latest version:
```
find_package(Libint 1.2.1 CONFIG QUIET COMPONENTS ${MAX_AM_ERI})
```
but it fails to find it:
```
-- Suitable Libint could not be located, Building Libint instead.
```
libint-1.2.1_1 doesn't provide .pc or .cmake files, so I am not sure how can it even ever be able to find it through a cmake script like this?
It's undesirable to download and rebuild libint when the package is available.
Please also note that the latest generation of libint is libint2, which does provide ```libint2.pc``` .
|
build
|
fails to find libint it doesn t provide cmake or pc files external upstream libint cmakelists txt has this line find package libint config quiet components max am eri i have the latest version of libint installed libint also see here i changed this line to the latest version find package libint config quiet components max am eri but it fails to find it suitable libint could not be located building libint instead libint doesn t provide pc or cmake files so i am not sure how can it even ever be able to find it through a cmake script like this it s undesirable to download and rebuild libint when the package is available please also note that the latest generation of libint is which does provide pc
| 1
|
61,892
| 15,102,719,908
|
IssuesEvent
|
2021-02-08 09:22:51
|
curl/curl
|
https://api.github.com/repos/curl/curl
|
closed
|
unresolved external symbol Visual Studio 2019 Community Edition, C++
|
Windows build
|
Hello Everyone,
I was using the pre-compiled version of CURL it was working fine, but I needed to enable SFTP so I downloaded from this repo, ran the bat file then compiled the CURL using x64 Native Tools Command Prompt for VS 2019 with the command:
`nmake /f Makefile.vc mode=static MACHINE=x64 WITH_SSL=static SSL_PATH=..\openssl-x64 WITH_SSH2=static SSH2_PATH=..\libssh2-x64
`
the compilation was fine without any errors, I then copied the folders bin, include, lib to my project folder (previously pre-compiled libcurl) and then I get errors for:
curl_global_init
curl_easy_init
curl_easy_setopt
curl_easy_perform
curl_easy_cleanup
curl_global_cleanup
meanwhile running curl.exe works as a charm!
Errors can be seen:
https://i.imgur.com/OHMmFOQ.jpg
### I did this
I tried different combination of compiling the libcur but to no avail
### I expected the following
I was expecting to be able to connect to SFTP with C++ app
### curl/libcurl version
```
curl.exe -V
curl 7.74.0-DEV (x86_64-pc-win32) libcurl/7.74.0-DEV OpenSSL/1.1.1g WinIDN libssh2/1.9.0
Release-Date: [unreleased]
Protocols: dict file ftp ftps gopher http https imap imaps ldap ldaps mqtt pop3 pop3s rtsp scp sftp smb smbs smtp smtps telnet tftp
Features: AsynchDNS HTTPS-proxy IDN IPv6 Kerberos Largefile NTLM SPNEGO SSL SSPI UnixSockets alt-svc
```
### operating system
Windows 10 x64
|
1.0
|
unresolved external symbol Visual Studio 2019 Community Edition, C++ - Hello Everyone,
I was using the pre-compiled version of CURL it was working fine, but I needed to enable SFTP so I downloaded from this repo, ran the bat file then compiled the CURL using x64 Native Tools Command Prompt for VS 2019 with the command:
`nmake /f Makefile.vc mode=static MACHINE=x64 WITH_SSL=static SSL_PATH=..\openssl-x64 WITH_SSH2=static SSH2_PATH=..\libssh2-x64
`
the compilation was fine without any errors, I then copied the folders bin, include, lib to my project folder (previously pre-compiled libcurl) and then I get errors for:
curl_global_init
curl_easy_init
curl_easy_setopt
curl_easy_perform
curl_easy_cleanup
curl_global_cleanup
meanwhile running curl.exe works as a charm!
Errors can be seen:
https://i.imgur.com/OHMmFOQ.jpg
### I did this
I tried different combination of compiling the libcur but to no avail
### I expected the following
I was expecting to be able to connect to SFTP with C++ app
### curl/libcurl version
```
curl.exe -V
curl 7.74.0-DEV (x86_64-pc-win32) libcurl/7.74.0-DEV OpenSSL/1.1.1g WinIDN libssh2/1.9.0
Release-Date: [unreleased]
Protocols: dict file ftp ftps gopher http https imap imaps ldap ldaps mqtt pop3 pop3s rtsp scp sftp smb smbs smtp smtps telnet tftp
Features: AsynchDNS HTTPS-proxy IDN IPv6 Kerberos Largefile NTLM SPNEGO SSL SSPI UnixSockets alt-svc
```
### operating system
Windows 10 x64
|
build
|
unresolved external symbol visual studio community edition c hello everyone i was using the pre compiled version of curl it was working fine but i needed to enable sftp so i downloaded from this repo ran the bat file then compiled the curl using native tools command prompt for vs with the command nmake f makefile vc mode static machine with ssl static ssl path openssl with static path the compilation was fine without any errors i then copied the folders bin include lib to my project folder previously pre compiled libcurl and then i get errors for curl global init curl easy init curl easy setopt curl easy perform curl easy cleanup curl global cleanup meanwhile running curl exe works as a charm errors can be seen i did this i tried different combination of compiling the libcur but to no avail i expected the following i was expecting to be able to connect to sftp with c app curl libcurl version curl exe v curl dev pc libcurl dev openssl winidn release date protocols dict file ftp ftps gopher http https imap imaps ldap ldaps mqtt rtsp scp sftp smb smbs smtp smtps telnet tftp features asynchdns https proxy idn kerberos largefile ntlm spnego ssl sspi unixsockets alt svc operating system windows
| 1
|
494,085
| 14,244,906,976
|
IssuesEvent
|
2020-11-19 07:48:02
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
There's no recaptcha on register by invitation page
|
Category: Accounts Priority: Critical Status: Fixed
|
Which leads to a failed validation for ALL invited customers. Production.
Also correct validation error message as it's "Username field has incorrect characters" even for re-captcha error.
|
1.0
|
There's no recaptcha on register by invitation page - Which leads to a failed validation for ALL invited customers. Production.
Also correct validation error message as it's "Username field has incorrect characters" even for re-captcha error.
|
non_build
|
there s no recaptcha on register by invitation page which leads to a failed validation for all invited customers production also correct validation error message as it s username field has incorrect characters even for re captcha error
| 0
|
20,044
| 6,808,397,393
|
IssuesEvent
|
2017-11-04 02:08:22
|
facebook/react
|
https://api.github.com/repos/facebook/react
|
opened
|
Investigate DCE regression
|
Component: Build Infrastructure
|
Seems like we started bundling persistent reconciler in DOM bundles. Maybe this negated some effect from ES modules (even though they were a net win). Need to look into why this happened.
|
1.0
|
Investigate DCE regression - Seems like we started bundling persistent reconciler in DOM bundles. Maybe this negated some effect from ES modules (even though they were a net win). Need to look into why this happened.
|
build
|
investigate dce regression seems like we started bundling persistent reconciler in dom bundles maybe this negated some effect from es modules even though they were a net win need to look into why this happened
| 1
|
25,320
| 2,679,202,896
|
IssuesEvent
|
2015-03-26 15:33:34
|
learnweb/moodle-mod_ratingallocate
|
https://api.github.com/repos/learnweb/moodle-mod_ratingallocate
|
closed
|
UI 'Rating is over'
|
Effort: High Priority: Very High
|
The UI Layout of the site after the rating period should be redesigned.
It is very important to have the buttons in the 'Distribution Table" section ordered properly. Maybe also the 'Start Distribution'-Button should be highlighted, so it is visable as 'the next step'.
Right now its just a bunch of salad if you see it for the first time :D
|
1.0
|
UI 'Rating is over' - The UI Layout of the site after the rating period should be redesigned.
It is very important to have the buttons in the 'Distribution Table" section ordered properly. Maybe also the 'Start Distribution'-Button should be highlighted, so it is visable as 'the next step'.
Right now its just a bunch of salad if you see it for the first time :D
|
non_build
|
ui rating is over the ui layout of the site after the rating period should be redesigned it is very important to have the buttons in the distribution table section ordered properly maybe also the start distribution button should be highlighted so it is visable as the next step right now its just a bunch of salad if you see it for the first time d
| 0
|
63,370
| 15,590,011,699
|
IssuesEvent
|
2021-03-18 08:49:59
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
opened
|
Revise text across screens
|
Bug P2 Study builder
|
Standardize the use of title case and lower case across screens, and update text accordingly.
for example, Add/Edit/View eligibility test, Add/Edit/View comprehension test and etc....
[Note : It should not apply for the title of the main screen Eg. Eligibility, Comprehension test, Questionnaires and etc..]
|
1.0
|
Revise text across screens - Standardize the use of title case and lower case across screens, and update text accordingly.
for example, Add/Edit/View eligibility test, Add/Edit/View comprehension test and etc....
[Note : It should not apply for the title of the main screen Eg. Eligibility, Comprehension test, Questionnaires and etc..]
|
build
|
revise text across screens standardize the use of title case and lower case across screens and update text accordingly for example add edit view eligibility test add edit view comprehension test and etc
| 1
|
4,628
| 3,411,861,237
|
IssuesEvent
|
2015-12-05 12:06:55
|
godotengine/godot
|
https://api.github.com/repos/godotengine/godot
|
opened
|
New webp driver does not build on linux with gcc 4.9.2
|
bug platform:linux topic:buildsystem
|
The error is:
```
drivers/webp/enc/webpenc.c: In function 'WebPEncode':
drivers/webp/enc/webpenc.c:324:14: error: 'WebPConfig' has no member named 'exact'
if (!config->exact) {
^
```
|
1.0
|
New webp driver does not build on linux with gcc 4.9.2 - The error is:
```
drivers/webp/enc/webpenc.c: In function 'WebPEncode':
drivers/webp/enc/webpenc.c:324:14: error: 'WebPConfig' has no member named 'exact'
if (!config->exact) {
^
```
|
build
|
new webp driver does not build on linux with gcc the error is drivers webp enc webpenc c in function webpencode drivers webp enc webpenc c error webpconfig has no member named exact if config exact
| 1
|
2,062
| 2,692,888,780
|
IssuesEvent
|
2015-04-01 11:18:00
|
wuic/wuic
|
https://api.github.com/repos/wuic/wuic
|
opened
|
[RequestDispatcher] Contextualized stream access
|
0.5.2 Code Doc
|
See: https://github.com/eclipse/jetty.project/commit/c23f21c761ff56f2ad4841a98083b237494442de#commitcomment-10515800
Application servers don't have to allow RequestDispatcher usage out of any HttpServletRequest scope since it's not required by specs. This means that we need to expose the original request when RequestDispatcher.include needs to be called.
Moreover, this will only works when an real HTTP request is processed. However, some servers like tomcat or more recently undertow allow to use RequestDispatcherDao during warm up or polling operation. The support of this feature for RequestDispatcherDao should be kept, execution context should be checked and documented since this is not crossplatform.
|
1.0
|
[RequestDispatcher] Contextualized stream access - See: https://github.com/eclipse/jetty.project/commit/c23f21c761ff56f2ad4841a98083b237494442de#commitcomment-10515800
Application servers don't have to allow RequestDispatcher usage out of any HttpServletRequest scope since it's not required by specs. This means that we need to expose the original request when RequestDispatcher.include needs to be called.
Moreover, this will only works when an real HTTP request is processed. However, some servers like tomcat or more recently undertow allow to use RequestDispatcherDao during warm up or polling operation. The support of this feature for RequestDispatcherDao should be kept, execution context should be checked and documented since this is not crossplatform.
|
non_build
|
contextualized stream access see application servers don t have to allow requestdispatcher usage out of any httpservletrequest scope since it s not required by specs this means that we need to expose the original request when requestdispatcher include needs to be called moreover this will only works when an real http request is processed however some servers like tomcat or more recently undertow allow to use requestdispatcherdao during warm up or polling operation the support of this feature for requestdispatcherdao should be kept execution context should be checked and documented since this is not crossplatform
| 0
|
1,551
| 2,780,826,685
|
IssuesEvent
|
2015-05-06 08:46:49
|
AppGyver/steroids
|
https://api.github.com/repos/AppGyver/steroids
|
closed
|
Build service does not sign App store builds, can't release new versions
|
Build Service investigating
|
Hi,
Application loader gives me this error message when I try to upload a new build to App store:
ERROR ITMS-90035: “Invalid Signature. Code object is not signed at all. The binary at path [myapp.app/Application/build/components/moment/scripts/npm_prepublish.sh] contains an invalid signature. Make sure you have signed your application with a distribution certificate, not an ad hoc certificate or a development certificate. Verify that the code signing settings in Xcode are correct at the target level (which override any values at the project level). Additionally, make sure the bundle you are uploading was built using a Release target in Xcode, not a Simulator target. If you are certain your code signing settings are correct, choose ”Clean All“ in Xcode, delete the ”build“ directory in the Finder, and rebuild your release target. For more information, please consult developer.apple.com/…/Introduction.html”
I have released a lot of successful builds of this app before, but since yesterday (22.4.) Application loader refuses to upload the app to App store because the code is not signed.
My distribution certificate was not expired, but I tried creating a new certificate anyway. Still the Application loader gives me the same error.
Is something broken in the build service or is this caused by something else?
|
1.0
|
Build service does not sign App store builds, can't release new versions - Hi,
Application loader gives me this error message when I try to upload a new build to App store:
ERROR ITMS-90035: “Invalid Signature. Code object is not signed at all. The binary at path [myapp.app/Application/build/components/moment/scripts/npm_prepublish.sh] contains an invalid signature. Make sure you have signed your application with a distribution certificate, not an ad hoc certificate or a development certificate. Verify that the code signing settings in Xcode are correct at the target level (which override any values at the project level). Additionally, make sure the bundle you are uploading was built using a Release target in Xcode, not a Simulator target. If you are certain your code signing settings are correct, choose ”Clean All“ in Xcode, delete the ”build“ directory in the Finder, and rebuild your release target. For more information, please consult developer.apple.com/…/Introduction.html”
I have released a lot of successful builds of this app before, but since yesterday (22.4.) Application loader refuses to upload the app to App store because the code is not signed.
My distribution certificate was not expired, but I tried creating a new certificate anyway. Still the Application loader gives me the same error.
Is something broken in the build service or is this caused by something else?
|
build
|
build service does not sign app store builds can t release new versions hi application loader gives me this error message when i try to upload a new build to app store error itms “invalid signature code object is not signed at all the binary at path contains an invalid signature make sure you have signed your application with a distribution certificate not an ad hoc certificate or a development certificate verify that the code signing settings in xcode are correct at the target level which override any values at the project level additionally make sure the bundle you are uploading was built using a release target in xcode not a simulator target if you are certain your code signing settings are correct choose ”clean all“ in xcode delete the ”build“ directory in the finder and rebuild your release target for more information please consult developer apple com … introduction html” i have released a lot of successful builds of this app before but since yesterday application loader refuses to upload the app to app store because the code is not signed my distribution certificate was not expired but i tried creating a new certificate anyway still the application loader gives me the same error is something broken in the build service or is this caused by something else
| 1
|
89,032
| 25,568,132,432
|
IssuesEvent
|
2022-11-30 15:40:50
|
getll/pokemon-web-services
|
https://api.github.com/repos/getll/pokemon-web-services
|
closed
|
TASK 5.b: Create POST methods
|
Build #2 Sub Task
|
Creating all POST callback functions for the following resources:
- /generations/{generationId}/games
- /trainers/{trainerId}/pokedex
- /pokemon/{pokemonId}/pokedex
### Requirements
- PHP IDE (NetBeans / VS Code)
- SLIM Framework
- Thunder Client (VS Code extension)
- XAMPP
- phpMyAdmin
### Dependencies
- Task 2
- Task 4
|
1.0
|
TASK 5.b: Create POST methods - Creating all POST callback functions for the following resources:
- /generations/{generationId}/games
- /trainers/{trainerId}/pokedex
- /pokemon/{pokemonId}/pokedex
### Requirements
- PHP IDE (NetBeans / VS Code)
- SLIM Framework
- Thunder Client (VS Code extension)
- XAMPP
- phpMyAdmin
### Dependencies
- Task 2
- Task 4
|
build
|
task b create post methods creating all post callback functions for the following resources generations generationid games trainers trainerid pokedex pokemon pokemonid pokedex requirements php ide netbeans vs code slim framework thunder client vs code extension xampp phpmyadmin dependencies task task
| 1
|
68,917
| 29,928,025,317
|
IssuesEvent
|
2023-06-22 07:27:13
|
amplication/amplication
|
https://api.github.com/repos/amplication/amplication
|
closed
|
🐛 Bug Report: Disconnect an organization is not reflected in the organizations selector
|
type: bug good first issue @amplication/git-service open to community
|
### What happened?
In sync with GitHub page, when a user disconnects an organization, it is not reflected in the selector

### What you expected to happen
I expect that when a user disconnects an organization, it will not appear in the selectordisconnects
### How to reproduce
- go to sync with GitHub page
- if you didn't sync before, connect your account/organization
- try to connect a new organization, but this time, click on `configure`

at the bottom of the page, you will see disconnect, click on that
- go back to Amplication, and see that your organization still appears in the selector
### Amplication version
0.15.0
### Environment
_No response_
### Are you willing to submit PR?
_No response_
|
1.0
|
🐛 Bug Report: Disconnect an organization is not reflected in the organizations selector - ### What happened?
In sync with GitHub page, when a user disconnects an organization, it is not reflected in the selector

### What you expected to happen
I expect that when a user disconnects an organization, it will not appear in the selectordisconnects
### How to reproduce
- go to sync with GitHub page
- if you didn't sync before, connect your account/organization
- try to connect a new organization, but this time, click on `configure`

at the bottom of the page, you will see disconnect, click on that
- go back to Amplication, and see that your organization still appears in the selector
### Amplication version
0.15.0
### Environment
_No response_
### Are you willing to submit PR?
_No response_
|
non_build
|
🐛 bug report disconnect an organization is not reflected in the organizations selector what happened in sync with github page when a user disconnects an organization it is not reflected in the selector what you expected to happen i expect that when a user disconnects an organization it will not appear in the selectordisconnects how to reproduce go to sync with github page if you didn t sync before connect your account organization try to connect a new organization but this time click on configure at the bottom of the page you will see disconnect click on that go back to amplication and see that your organization still appears in the selector amplication version environment no response are you willing to submit pr no response
| 0
|
188,469
| 14,445,662,383
|
IssuesEvent
|
2020-12-07 23:27:20
|
wsu-cpts489-fa20/bp-kliks
|
https://api.github.com/repos/wsu-cpts489-fa20/bp-kliks
|
closed
|
Tests for receiving feedback after submitting a response - Student
|
Story Points: 1 Tests
|
**Overview**:
Write the tests for being view a message after submitting a response as a student.
**Acceptance Criteria**:
- [ ] Tests to view a message of success or failure after submitting a response.
**Connected Issue(s)**:
Tests for issue #5
|
1.0
|
Tests for receiving feedback after submitting a response - Student - **Overview**:
Write the tests for being view a message after submitting a response as a student.
**Acceptance Criteria**:
- [ ] Tests to view a message of success or failure after submitting a response.
**Connected Issue(s)**:
Tests for issue #5
|
non_build
|
tests for receiving feedback after submitting a response student overview write the tests for being view a message after submitting a response as a student acceptance criteria tests to view a message of success or failure after submitting a response connected issue s tests for issue
| 0
|
735,460
| 25,399,382,426
|
IssuesEvent
|
2022-11-22 10:55:26
|
TalaoDAO/AltMe
|
https://api.github.com/repos/TalaoDAO/AltMe
|
closed
|
Create Ethereum account during onbording
|
a V3 Priority ethereum
|
The same way we are creating tezos account during onboard (create and import), we are creating the first ethereum account
|
1.0
|
Create Ethereum account during onbording - The same way we are creating tezos account during onboard (create and import), we are creating the first ethereum account
|
non_build
|
create ethereum account during onbording the same way we are creating tezos account during onboard create and import we are creating the first ethereum account
| 0
|
20,475
| 6,892,119,418
|
IssuesEvent
|
2017-11-22 19:39:22
|
moby/moby
|
https://api.github.com/repos/moby/moby
|
closed
|
`docker commit --change 'ENV X=...$X'` does not substitute image environment
|
area/builder version/17.09
|
## Description
I expect the following two things to be _roughly_ equivalent, however they are not. My actual usecase involves `$PATH` augmentation and a more complex `commit` cycle, the "example" in this issue is a MVCE. I've been careful in the example to quote properly -- this is not a substitution coming from the shell running `docker commit`.
### via dockerfile
```dockerfile
FROM ubuntu:xenial
ENV X=1
ENV X=2:$X
```
```bash
docker build -q -t test .
docker inspect test --format='{{.Config.Env}}'
```
```console
$ docker build -q -t test .
sha256:9e1a10d984e0e2457620c83175888546820272b82fb0103d65da273ab74dd732
$ docker inspect test --format='{{.Config.Env}}'
[PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin X=2:1]
```
**Notice here that `X=2:1`**
### via `docker commit`
```dockerfile
FROM ubuntu:xenial
ENV X=1
```
```bash
docker build -q -t test .
CID=$(docker create test)
docker commit --change 'ENV X=2:$X' "$CID" test
docker rm "$CID"
docker inspect test --format='{{.Config.Env}}'
```
```console
$ docker build -q -t test .
sha256:2d7a6b4f468a894a38a669f8c9a826cac17f2f6eb4ff7f1a7bc51d6e5950325a
$ CID=$(docker create test)
$ docker commit --change 'ENV X=2:$X' "$CID" test
sha256:7f9767be8464c4a48ad5c7545f7986c5b8a898ab1b64026ecaad2dfb0409146a
$ docker rm "$CID"
8c7957c982bf4b34eb91d5268c9cfe96d2b49743b9342850c3d9d99f504996f2
$ docker inspect test --format='{{.Config.Env}}'
[X=2: PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin]
```
**Notice here that `X=2:`** (the environment substitution from the parent is lost)
**Output of `docker version`:**
```
$ docker version
Client:
Version: 17.09.0-ce
API version: 1.32
Go version: go1.8.3
Git commit: afdb6d4
Built: Tue Sep 26 22:42:18 2017
OS/Arch: linux/amd64
Server:
Version: 17.09.0-ce
API version: 1.32 (minimum version 1.12)
Go version: go1.8.3
Git commit: afdb6d4
Built: Tue Sep 26 22:40:56 2017
OS/Arch: linux/amd64
Experimental: false
```
**Output of `docker info`:**
```
$ docker info
Containers: 0
Running: 0
Paused: 0
Stopped: 0
Images: 35
Server Version: 17.09.0-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 06b9cb35161009dcb7123345749fef02f7cea8e0
runc version: 3f2f8b84a77f73d38244dd690525642a72156c64
init version: 949e6fa
Security Options:
apparmor
seccomp
Profile: default
Kernel Version: 4.10.0-40-generic
Operating System: Ubuntu 16.04.3 LTS
OSType: linux
Architecture: x86_64
CPUs: 1
Total Memory: 7.795GiB
Name: asottile-VirtualBox
ID: MCDD:BXAL:Y4PM:SIA7:I73R:W3F6:C4IY:ZEP4:K44W:XLYA:USI2:FAMT
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
WARNING: No swap limit support
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
virtualbox, but I don't think it's important.
|
1.0
|
`docker commit --change 'ENV X=...$X'` does not substitute image environment - ## Description
I expect the following two things to be _roughly_ equivalent, however they are not. My actual usecase involves `$PATH` augmentation and a more complex `commit` cycle, the "example" in this issue is a MVCE. I've been careful in the example to quote properly -- this is not a substitution coming from the shell running `docker commit`.
### via dockerfile
```dockerfile
FROM ubuntu:xenial
ENV X=1
ENV X=2:$X
```
```bash
docker build -q -t test .
docker inspect test --format='{{.Config.Env}}'
```
```console
$ docker build -q -t test .
sha256:9e1a10d984e0e2457620c83175888546820272b82fb0103d65da273ab74dd732
$ docker inspect test --format='{{.Config.Env}}'
[PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin X=2:1]
```
**Notice here that `X=2:1`**
### via `docker commit`
```dockerfile
FROM ubuntu:xenial
ENV X=1
```
```bash
docker build -q -t test .
CID=$(docker create test)
docker commit --change 'ENV X=2:$X' "$CID" test
docker rm "$CID"
docker inspect test --format='{{.Config.Env}}'
```
```console
$ docker build -q -t test .
sha256:2d7a6b4f468a894a38a669f8c9a826cac17f2f6eb4ff7f1a7bc51d6e5950325a
$ CID=$(docker create test)
$ docker commit --change 'ENV X=2:$X' "$CID" test
sha256:7f9767be8464c4a48ad5c7545f7986c5b8a898ab1b64026ecaad2dfb0409146a
$ docker rm "$CID"
8c7957c982bf4b34eb91d5268c9cfe96d2b49743b9342850c3d9d99f504996f2
$ docker inspect test --format='{{.Config.Env}}'
[X=2: PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin]
```
**Notice here that `X=2:`** (the environment substitution from the parent is lost)
**Output of `docker version`:**
```
$ docker version
Client:
Version: 17.09.0-ce
API version: 1.32
Go version: go1.8.3
Git commit: afdb6d4
Built: Tue Sep 26 22:42:18 2017
OS/Arch: linux/amd64
Server:
Version: 17.09.0-ce
API version: 1.32 (minimum version 1.12)
Go version: go1.8.3
Git commit: afdb6d4
Built: Tue Sep 26 22:40:56 2017
OS/Arch: linux/amd64
Experimental: false
```
**Output of `docker info`:**
```
$ docker info
Containers: 0
Running: 0
Paused: 0
Stopped: 0
Images: 35
Server Version: 17.09.0-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 06b9cb35161009dcb7123345749fef02f7cea8e0
runc version: 3f2f8b84a77f73d38244dd690525642a72156c64
init version: 949e6fa
Security Options:
apparmor
seccomp
Profile: default
Kernel Version: 4.10.0-40-generic
Operating System: Ubuntu 16.04.3 LTS
OSType: linux
Architecture: x86_64
CPUs: 1
Total Memory: 7.795GiB
Name: asottile-VirtualBox
ID: MCDD:BXAL:Y4PM:SIA7:I73R:W3F6:C4IY:ZEP4:K44W:XLYA:USI2:FAMT
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
WARNING: No swap limit support
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
virtualbox, but I don't think it's important.
|
build
|
docker commit change env x x does not substitute image environment description i expect the following two things to be roughly equivalent however they are not my actual usecase involves path augmentation and a more complex commit cycle the example in this issue is a mvce i ve been careful in the example to quote properly this is not a substitution coming from the shell running docker commit via dockerfile dockerfile from ubuntu xenial env x env x x bash docker build q t test docker inspect test format config env console docker build q t test docker inspect test format config env notice here that x via docker commit dockerfile from ubuntu xenial env x bash docker build q t test cid docker create test docker commit change env x x cid test docker rm cid docker inspect test format config env console docker build q t test cid docker create test docker commit change env x x cid test docker rm cid docker inspect test format config env notice here that x the environment substitution from the parent is lost output of docker version docker version client version ce api version go version git commit built tue sep os arch linux server version ce api version minimum version go version git commit built tue sep os arch linux experimental false output of docker info docker info containers running paused stopped images server version ce storage driver backing filesystem extfs supports d type true native overlay diff true logging driver json file cgroup driver cgroupfs plugins volume local network bridge host macvlan null overlay log awslogs fluentd gcplogs gelf journald json file logentries splunk syslog swarm inactive runtimes runc default runtime runc init binary docker init containerd version runc version init version security options apparmor seccomp profile default kernel version generic operating system ubuntu lts ostype linux architecture cpus total memory name asottile virtualbox id mcdd bxal xlya famt docker root dir var lib docker debug mode client false debug mode server false registry experimental false insecure registries live restore enabled false warning no swap limit support additional environment details aws virtualbox physical etc virtualbox but i don t think it s important
| 1
|
58,566
| 7,163,645,410
|
IssuesEvent
|
2018-01-29 08:22:06
|
buzzn/scrum
|
https://api.github.com/repos/buzzn/scrum
|
closed
|
As Pho I want to import and see all the prices belonging to a group.
|
Design needed User Story Wireframe needed estimation size: 3
|
**PR**
---
**Story**
As Pho I want to import and see all the prices belonging to a group
---
**XD**
- [ ] check designs
**DoD**
- [ ] show current prices
- [ ] make sure that
exclude:
- price history
|
1.0
|
As Pho I want to import and see all the prices belonging to a group. - **PR**
---
**Story**
As Pho I want to import and see all the prices belonging to a group
---
**XD**
- [ ] check designs
**DoD**
- [ ] show current prices
- [ ] make sure that
exclude:
- price history
|
non_build
|
as pho i want to import and see all the prices belonging to a group pr story as pho i want to import and see all the prices belonging to a group xd check designs dod show current prices make sure that exclude price history
| 0
|
664,279
| 22,264,039,729
|
IssuesEvent
|
2022-06-10 05:14:34
|
wso2/api-manager
|
https://api.github.com/repos/wso2/api-manager
|
opened
|
Multi Endpoint Configuration from api_params.yaml
|
Type/New Feature Priority/Normal
|
Description:
When importing an API using the API Controller, there is a requirement of mentioning multiple endpoints in the api_params.yaml file to handle the load balancing and failover scenarios with relevant configurations, as the publisher portal already supports those.
Suggested Assignees:
@BLasan
Affected Product Version:
APICTL 3.1
Related Issues:
[product-apim-tooling-917](https://github.com/wso2/product-apim-tooling/pull/917)
|
1.0
|
Multi Endpoint Configuration from api_params.yaml - Description:
When importing an API using the API Controller, there is a requirement of mentioning multiple endpoints in the api_params.yaml file to handle the load balancing and failover scenarios with relevant configurations, as the publisher portal already supports those.
Suggested Assignees:
@BLasan
Affected Product Version:
APICTL 3.1
Related Issues:
[product-apim-tooling-917](https://github.com/wso2/product-apim-tooling/pull/917)
|
non_build
|
multi endpoint configuration from api params yaml description when importing an api using the api controller there is a requirement of mentioning multiple endpoints in the api params yaml file to handle the load balancing and failover scenarios with relevant configurations as the publisher portal already supports those suggested assignees blasan affected product version apictl related issues
| 0
|
826,074
| 31,551,021,247
|
IssuesEvent
|
2023-09-02 04:02:37
|
space-wizards/space-station-14
|
https://api.github.com/repos/space-wizards/space-station-14
|
closed
|
Trash bag bug
|
Issue: Bug Priority: 2-Before Release Size: Very Small Difficulty: 2-Medium
|
## Description
The trash bag dumps trash onto the ground instead of the disposals' bin.
**Reproduction**
Spawn into the game and get yourself a trash bag. Find trash. Pick up any trash from the ground. Either LMB on the disposals' unit directly or RMB on it and press "dump out into disposal unit".
**Screenshots**

**Additional context**
Yesterday I saw that the space cleaner is broken too. It would be wise to fix these issues quickly, as these are one of the main job focuses of a janitor.
|
1.0
|
Trash bag bug - ## Description
The trash bag dumps trash onto the ground instead of the disposals' bin.
**Reproduction**
Spawn into the game and get yourself a trash bag. Find trash. Pick up any trash from the ground. Either LMB on the disposals' unit directly or RMB on it and press "dump out into disposal unit".
**Screenshots**

**Additional context**
Yesterday I saw that the space cleaner is broken too. It would be wise to fix these issues quickly, as these are one of the main job focuses of a janitor.
|
non_build
|
trash bag bug description the trash bag dumps trash onto the ground instead of the disposals bin reproduction spawn into the game and get yourself a trash bag find trash pick up any trash from the ground either lmb on the disposals unit directly or rmb on it and press dump out into disposal unit screenshots additional context yesterday i saw that the space cleaner is broken too it would be wise to fix these issues quickly as these are one of the main job focuses of a janitor
| 0
|
7,615
| 4,020,461,369
|
IssuesEvent
|
2016-05-16 18:30:00
|
mitchellh/packer
|
https://api.github.com/repos/mitchellh/packer
|
opened
|
Azure: Where's the Best Error Message
|
builder/azure
|
A customer experienced an issue when the *capture_name_prefix* was set to an unacceptable value (#3535). The method signature for the API call returns an HTTP response and an error. The Azure builder code discards the HTTP response, and checks the error only. The error's message is not intuitive, and does not return any information to help to debug the issue. The HTTP response is more helpful, but it is discarded.
The fix is for the builder code to check and surface both. (Checking the error only is sufficient to indicate there is an error with the API call.) I am tracking the [issue](https://github.com/Azure/azure-sdk-for-go/issues/328) with the Azure SDK team too, to see what their recommendations are.
|
1.0
|
Azure: Where's the Best Error Message - A customer experienced an issue when the *capture_name_prefix* was set to an unacceptable value (#3535). The method signature for the API call returns an HTTP response and an error. The Azure builder code discards the HTTP response, and checks the error only. The error's message is not intuitive, and does not return any information to help to debug the issue. The HTTP response is more helpful, but it is discarded.
The fix is for the builder code to check and surface both. (Checking the error only is sufficient to indicate there is an error with the API call.) I am tracking the [issue](https://github.com/Azure/azure-sdk-for-go/issues/328) with the Azure SDK team too, to see what their recommendations are.
|
build
|
azure where s the best error message a customer experienced an issue when the capture name prefix was set to an unacceptable value the method signature for the api call returns an http response and an error the azure builder code discards the http response and checks the error only the error s message is not intuitive and does not return any information to help to debug the issue the http response is more helpful but it is discarded the fix is for the builder code to check and surface both checking the error only is sufficient to indicate there is an error with the api call i am tracking the with the azure sdk team too to see what their recommendations are
| 1
|
271,282
| 23,593,578,052
|
IssuesEvent
|
2022-08-23 17:12:04
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Itapeva
|
generalization test development template - Betha tag - Obras Públicas subtag - Dados para acompanhamento
|
DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Itapeva.
|
1.0
|
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Itapeva - DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Itapeva.
|
non_build
|
teste de generalizacao para a tag obras públicas dados para acompanhamento itapeva dod realizar o teste de generalização do validador da tag obras públicas dados para acompanhamento para o município de itapeva
| 0
|
32,812
| 8,956,270,883
|
IssuesEvent
|
2019-01-26 16:08:28
|
vavr-io/vavr
|
https://api.github.com/repos/vavr-io/vavr
|
opened
|
vavr-match should only be a compile time dependency of vavr
|
bug build/infrastructure/ide
|
The vavr-match dependency 'leaked' into the library with the PR #2294.
We need to revert that.

This should not be a backward compatibility problem.
Migration to the fix: add vavr-match as compile time dependency to the actual application build.
|
1.0
|
vavr-match should only be a compile time dependency of vavr - The vavr-match dependency 'leaked' into the library with the PR #2294.
We need to revert that.

This should not be a backward compatibility problem.
Migration to the fix: add vavr-match as compile time dependency to the actual application build.
|
build
|
vavr match should only be a compile time dependency of vavr the vavr match dependency leaked into the library with the pr we need to revert that this should not be a backward compatibility problem migration to the fix add vavr match as compile time dependency to the actual application build
| 1
|
450,258
| 12,992,626,882
|
IssuesEvent
|
2020-07-23 07:17:51
|
returntocorp/semgrep
|
https://api.github.com/repos/returntocorp/semgrep
|
closed
|
Make message and severity fields optional
|
enhancement feature:patterns priority:low
|
At the moment, the ocaml-core errors if it doesn't get a message or severity field when parsing YAML. No reason not to have defaults for these
|
1.0
|
Make message and severity fields optional - At the moment, the ocaml-core errors if it doesn't get a message or severity field when parsing YAML. No reason not to have defaults for these
|
non_build
|
make message and severity fields optional at the moment the ocaml core errors if it doesn t get a message or severity field when parsing yaml no reason not to have defaults for these
| 0
|
204,992
| 15,577,486,136
|
IssuesEvent
|
2021-03-17 13:35:32
|
AdoptOpenJDK/openjdk-infrastructure
|
https://api.github.com/repos/AdoptOpenJDK/openjdk-infrastructure
|
opened
|
Rename packet to equinix in hostnames
|
provider:packet testFail
|
Packet hosting became Equinix some time ago but we have not updated out naming convention to reflect that.
|
1.0
|
Rename packet to equinix in hostnames - Packet hosting became Equinix some time ago but we have not updated out naming convention to reflect that.
|
non_build
|
rename packet to equinix in hostnames packet hosting became equinix some time ago but we have not updated out naming convention to reflect that
| 0
|
68,852
| 17,438,829,720
|
IssuesEvent
|
2021-08-05 00:07:55
|
MicrosoftDocs/visualstudio-docs
|
https://api.github.com/repos/MicrosoftDocs/visualstudio-docs
|
closed
|
Ambiguity in overriding ToolsVersion settings
|
Pri2 auto-close doc-bug msbuild/tech stale visual-studio-dev15/prod
|
The "Order of Precedence" section should describe a 100% precise and unambiguous algorithm for determining the ToolsVersion to be used to build the given project.
Many of the cases mention extremely ambiguous terms "ToolsVersion" and "current ToolsVersion". The whole algorithm is about determining "the ToolsVersion value to use"; there is a "ToolsVersions" attribute in the project, "ToolsVersions" property in the MSBuild task, a "ToolsVersions" value of the version of Visual Studio used or installed on the machine. I have no idea which of these distinct things is used in each context in the algorithm description. the cases listed are not mutually exclusive and not exhaustive. Please apply strict logic to fix this algorithm and make it precise and unambiguous.
Copied from https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_queries/edit/460302/?triage=true
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 578f06ee-4e25-db51-aafc-0932c805b455
* Version Independent ID: a47ae76e-5748-b469-fce2-03e0e7f15fb7
* Content: [Overriding ToolsVersion Settings - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/msbuild/overriding-toolsversion-settings#order-of-precedence)
* Content Source: [docs/msbuild/overriding-toolsversion-settings.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/msbuild/overriding-toolsversion-settings.md)
* Product: **visual-studio-dev15**
* GitHub Login: @Mikejo5000
* Microsoft Alias: **mikejo**
|
1.0
|
Ambiguity in overriding ToolsVersion settings - The "Order of Precedence" section should describe a 100% precise and unambiguous algorithm for determining the ToolsVersion to be used to build the given project.
Many of the cases mention extremely ambiguous terms "ToolsVersion" and "current ToolsVersion". The whole algorithm is about determining "the ToolsVersion value to use"; there is a "ToolsVersions" attribute in the project, "ToolsVersions" property in the MSBuild task, a "ToolsVersions" value of the version of Visual Studio used or installed on the machine. I have no idea which of these distinct things is used in each context in the algorithm description. the cases listed are not mutually exclusive and not exhaustive. Please apply strict logic to fix this algorithm and make it precise and unambiguous.
Copied from https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_queries/edit/460302/?triage=true
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 578f06ee-4e25-db51-aafc-0932c805b455
* Version Independent ID: a47ae76e-5748-b469-fce2-03e0e7f15fb7
* Content: [Overriding ToolsVersion Settings - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/msbuild/overriding-toolsversion-settings#order-of-precedence)
* Content Source: [docs/msbuild/overriding-toolsversion-settings.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/msbuild/overriding-toolsversion-settings.md)
* Product: **visual-studio-dev15**
* GitHub Login: @Mikejo5000
* Microsoft Alias: **mikejo**
|
build
|
ambiguity in overriding toolsversion settings the order of precedence section should describe a precise and unambiguous algorithm for determining the toolsversion to be used to build the given project many of the cases mention extremely ambiguous terms toolsversion and current toolsversion the whole algorithm is about determining the toolsversion value to use there is a toolsversions attribute in the project toolsversions property in the msbuild task a toolsversions value of the version of visual studio used or installed on the machine i have no idea which of these distinct things is used in each context in the algorithm description the cases listed are not mutually exclusive and not exhaustive please apply strict logic to fix this algorithm and make it precise and unambiguous copied from document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id aafc version independent id content content source product visual studio github login microsoft alias mikejo
| 1
|
424,731
| 29,173,825,843
|
IssuesEvent
|
2023-05-19 05:57:15
|
RADB/Offord-EDI-2020
|
https://api.github.com/repos/RADB/Offord-EDI-2020
|
closed
|
Publish e-EDI as 64 bit application
|
documentation 2023
|
After building the application in 64 bit you must change the application pool to 64 bit
IIS -- Application Pools -- <APP POOL> -- Advanced Settings -- Enable 32 Bit Applications -- FALSE (64bit)

|
1.0
|
Publish e-EDI as 64 bit application - After building the application in 64 bit you must change the application pool to 64 bit
IIS -- Application Pools -- <APP POOL> -- Advanced Settings -- Enable 32 Bit Applications -- FALSE (64bit)

|
non_build
|
publish e edi as bit application after building the application in bit you must change the application pool to bit iis application pools advanced settings enable bit applications false
| 0
|
15,721
| 19,863,541,756
|
IssuesEvent
|
2022-01-22 06:37:18
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Automation Account Source Control Failure when Setting up VsoTfvc
|
automation/svc triaged cxp awaiting-product-team-response product-issue process-automation/subsvc Pri3
|
When trying to add a VsoTfvc repository to an automation account an error:
Create source control failed.
An error occurred while creating the source control named '[NameHere]'.
Error details:
The property 'branch' is required..
is generated and the instance is not created.
When selecting VsoTfvc the Branch dropdown is disabled with the text "Branch is not supported for VsoTfvc".
However I am able to successfully create the instance using the PowerShell New-AzAutomationSourceControl command.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
1.0
|
Automation Account Source Control Failure when Setting up VsoTfvc - When trying to add a VsoTfvc repository to an automation account an error:
Create source control failed.
An error occurred while creating the source control named '[NameHere]'.
Error details:
The property 'branch' is required..
is generated and the instance is not created.
When selecting VsoTfvc the Branch dropdown is disabled with the text "Branch is not supported for VsoTfvc".
However I am able to successfully create the instance using the PowerShell New-AzAutomationSourceControl command.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
non_build
|
automation account source control failure when setting up vsotfvc when trying to add a vsotfvc repository to an automation account an error create source control failed an error occurred while creating the source control named error details the property branch is required is generated and the instance is not created when selecting vsotfvc the branch dropdown is disabled with the text branch is not supported for vsotfvc however i am able to successfully create the instance using the powershell new azautomationsourcecontrol command document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
| 0
|
29,411
| 8,354,934,290
|
IssuesEvent
|
2018-10-02 14:31:37
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
openstack builder failing with 1.3.0
|
builder/openstack regression
|
openstack builder doesn't seem to work with 1.3.0:
```
openstack output will be in this color.
==> openstack: Loading flavor: generic2.small
openstack: Verified flavor. ID: 357a88f8-2af7-414c-a197-a2db9bda4e7b
==> openstack: Pausing after run of step 'StepLoadFlavor'. Press enter to continue.
==> openstack: Creating temporary keypair: packer_5b9ad626-cd24-d24c-a3bc-b5724a7523c1 ...
==> openstack: Created temporary keypair: packer_5b9ad626-cd24-d24c-a3bc-b5724a7523c1
==> openstack: Successfully converted BER encoded SSH key to DER encoding.
openstack: Saving key for debug purposes: os_openstack.pem
==> openstack: Pausing after run of step 'StepKeyPair'. Press enter to continue.
==> openstack: Pausing after run of step 'StepSourceImageInfo'. Press enter to continue.
==> openstack: Pausing after run of step 'StepCreateVolume'. Press enter to continue.
==> openstack: Launching server...
==> openstack: Launching server...
openstack: Server ID: 608f6674-d8a8-4355-9e28-174c1e5c46e6
==> openstack: Waiting for server to become ready...
==> openstack: Pausing after run of step 'StepRunSourceServer'. Press enter to continue.
==> openstack: Pausing after run of step 'StepGetPassword'. Press enter to continue.
==> openstack: Pausing after run of step 'StepWaitForRackConnect'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepWaitForRackConnect'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepGetPassword'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepRunSourceServer'. Press enter to continue.
==> openstack: Terminating the source server: 608f6674-d8a8-4355-9e28-174c1e5c46e6 ...
==> openstack: Pausing before cleanup of step 'StepCreateVolume'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepSourceImageInfo'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepKeyPair'. Press enter to continue.
==> openstack: Deleting temporary keypair: packer_5b9ad626-cd24-d24c-a3bc-b5724a7523c1 ...
==> openstack: Pausing before cleanup of step 'StepLoadFlavor'. Press enter to continue.
Build 'openstack' errored: unexpected EOF
==> Some builds didn't complete successfully and had errors:
--> openstack: unexpected EOF
```
|
1.0
|
openstack builder failing with 1.3.0 - openstack builder doesn't seem to work with 1.3.0:
```
openstack output will be in this color.
==> openstack: Loading flavor: generic2.small
openstack: Verified flavor. ID: 357a88f8-2af7-414c-a197-a2db9bda4e7b
==> openstack: Pausing after run of step 'StepLoadFlavor'. Press enter to continue.
==> openstack: Creating temporary keypair: packer_5b9ad626-cd24-d24c-a3bc-b5724a7523c1 ...
==> openstack: Created temporary keypair: packer_5b9ad626-cd24-d24c-a3bc-b5724a7523c1
==> openstack: Successfully converted BER encoded SSH key to DER encoding.
openstack: Saving key for debug purposes: os_openstack.pem
==> openstack: Pausing after run of step 'StepKeyPair'. Press enter to continue.
==> openstack: Pausing after run of step 'StepSourceImageInfo'. Press enter to continue.
==> openstack: Pausing after run of step 'StepCreateVolume'. Press enter to continue.
==> openstack: Launching server...
==> openstack: Launching server...
openstack: Server ID: 608f6674-d8a8-4355-9e28-174c1e5c46e6
==> openstack: Waiting for server to become ready...
==> openstack: Pausing after run of step 'StepRunSourceServer'. Press enter to continue.
==> openstack: Pausing after run of step 'StepGetPassword'. Press enter to continue.
==> openstack: Pausing after run of step 'StepWaitForRackConnect'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepWaitForRackConnect'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepGetPassword'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepRunSourceServer'. Press enter to continue.
==> openstack: Terminating the source server: 608f6674-d8a8-4355-9e28-174c1e5c46e6 ...
==> openstack: Pausing before cleanup of step 'StepCreateVolume'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepSourceImageInfo'. Press enter to continue.
==> openstack: Pausing before cleanup of step 'StepKeyPair'. Press enter to continue.
==> openstack: Deleting temporary keypair: packer_5b9ad626-cd24-d24c-a3bc-b5724a7523c1 ...
==> openstack: Pausing before cleanup of step 'StepLoadFlavor'. Press enter to continue.
Build 'openstack' errored: unexpected EOF
==> Some builds didn't complete successfully and had errors:
--> openstack: unexpected EOF
```
|
build
|
openstack builder failing with openstack builder doesn t seem to work with openstack output will be in this color openstack loading flavor small openstack verified flavor id openstack pausing after run of step steploadflavor press enter to continue openstack creating temporary keypair packer openstack created temporary keypair packer openstack successfully converted ber encoded ssh key to der encoding openstack saving key for debug purposes os openstack pem openstack pausing after run of step stepkeypair press enter to continue openstack pausing after run of step stepsourceimageinfo press enter to continue openstack pausing after run of step stepcreatevolume press enter to continue openstack launching server openstack launching server openstack server id openstack waiting for server to become ready openstack pausing after run of step steprunsourceserver press enter to continue openstack pausing after run of step stepgetpassword press enter to continue openstack pausing after run of step stepwaitforrackconnect press enter to continue openstack pausing before cleanup of step stepwaitforrackconnect press enter to continue openstack pausing before cleanup of step stepgetpassword press enter to continue openstack pausing before cleanup of step steprunsourceserver press enter to continue openstack terminating the source server openstack pausing before cleanup of step stepcreatevolume press enter to continue openstack pausing before cleanup of step stepsourceimageinfo press enter to continue openstack pausing before cleanup of step stepkeypair press enter to continue openstack deleting temporary keypair packer openstack pausing before cleanup of step steploadflavor press enter to continue build openstack errored unexpected eof some builds didn t complete successfully and had errors openstack unexpected eof
| 1
|
43,255
| 11,196,670,070
|
IssuesEvent
|
2020-01-03 10:47:10
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
closed
|
Include file when building tensorflow 2.0 from source missing.
|
subtype: ubuntu/linux type:build/install
|
When compiled, tensorflow2.0 outputs the libtensorflow_cc.so and libtensorflow_framework.so
Used: bazel build --config=opt //tensorflow:libtensorflow_cc.so
However there are no more genfiles where I was getting the include files to link to . In the tensorflow/cc/ops directory e.g. I am missing array_ops. There are many others missing. Before I could those from genfiles directory, now this folder is gone. Where can I get all the include files to compile my file with?
The alternative I think would be to use bazel build <my file>, however this takes too long to compile my file. I just want to use g++ to link with the above lib and the include files.
------------------------
OS: Linux Ubuntu 14.04
Tensorflow 2.0 (master) installed from source
No Cuda
C++ API
|
1.0
|
Include file when building tensorflow 2.0 from source missing. - When compiled, tensorflow2.0 outputs the libtensorflow_cc.so and libtensorflow_framework.so
Used: bazel build --config=opt //tensorflow:libtensorflow_cc.so
However there are no more genfiles where I was getting the include files to link to . In the tensorflow/cc/ops directory e.g. I am missing array_ops. There are many others missing. Before I could those from genfiles directory, now this folder is gone. Where can I get all the include files to compile my file with?
The alternative I think would be to use bazel build <my file>, however this takes too long to compile my file. I just want to use g++ to link with the above lib and the include files.
------------------------
OS: Linux Ubuntu 14.04
Tensorflow 2.0 (master) installed from source
No Cuda
C++ API
|
build
|
include file when building tensorflow from source missing when compiled outputs the libtensorflow cc so and libtensorflow framework so used bazel build config opt tensorflow libtensorflow cc so however there are no more genfiles where i was getting the include files to link to in the tensorflow cc ops directory e g i am missing array ops there are many others missing before i could those from genfiles directory now this folder is gone where can i get all the include files to compile my file with the alternative i think would be to use bazel build however this takes too long to compile my file i just want to use g to link with the above lib and the include files os linux ubuntu tensorflow master installed from source no cuda c api
| 1
|
259,904
| 8,201,462,680
|
IssuesEvent
|
2018-09-01 17:59:47
|
fossasia/open-event-webapp
|
https://api.github.com/repos/fossasia/open-event-webapp
|
opened
|
Tickets button not available
|
Priority: Urgent bug
|
The tickets button is no longer available when a new site is generated. Please reimplement as before.
|
1.0
|
Tickets button not available - The tickets button is no longer available when a new site is generated. Please reimplement as before.
|
non_build
|
tickets button not available the tickets button is no longer available when a new site is generated please reimplement as before
| 0
|
85,768
| 24,679,041,921
|
IssuesEvent
|
2022-10-18 19:33:43
|
hexops/mach
|
https://api.github.com/repos/hexops/mach
|
closed
|
eliminate symlinks from sdk-macos-12.0
|
build
|
https://github.com/hexops/sdk-macos-12.0 contains symlinks which are not reliable on Windows.
We should eliminate symlinks in that repo, similar to how we do for sdk-macos-11.3
|
1.0
|
eliminate symlinks from sdk-macos-12.0 - https://github.com/hexops/sdk-macos-12.0 contains symlinks which are not reliable on Windows.
We should eliminate symlinks in that repo, similar to how we do for sdk-macos-11.3
|
build
|
eliminate symlinks from sdk macos contains symlinks which are not reliable on windows we should eliminate symlinks in that repo similar to how we do for sdk macos
| 1
|
445,104
| 12,826,196,809
|
IssuesEvent
|
2020-07-06 16:08:43
|
osulp/Scholars-Archive
|
https://api.github.com/repos/osulp/Scholars-Archive
|
closed
|
Staging Cleanup: Admin permissions change?
|
Priority: Critical
|
I deposited into each worktype to test workflows. While the blue status “button” listed the correct stage of review (Grad School Review, Library Review, Pending Review, Honors College Review) and I could approve/advance as needed from the show page, none of the items showed up in the review queue for me.
After discussion with @wickr, I added myself as a Manager for each Admin Set. Then I could see all the queues.
I thought I was able to see items in the queue without being a Manager — there were even some Grad School Review items in my queue from March tests — since I was part of the Repo Managers group?
From wickr:
> this is from the Samvera docs: "In the Dashboard, click on “Tasks,” then “Review Submissions.” You will see a list of works requiring your review. Repository Admins will see all works in the repository, other reviewers will only see works in Admin Sets for which they are designated Managers."
So:
1. I am not sure what altered the Repo Managers group permissions and what impact it would have on production.
2. It looks like there are new Admin Sets for GTDs and HCTs? I don’t know how that is related to the above but if appropriate, it would be nice if old/unused ones were removed.
3. Whatever the situation, the goal is to have staging and prod match.
|
1.0
|
Staging Cleanup: Admin permissions change? - I deposited into each worktype to test workflows. While the blue status “button” listed the correct stage of review (Grad School Review, Library Review, Pending Review, Honors College Review) and I could approve/advance as needed from the show page, none of the items showed up in the review queue for me.
After discussion with @wickr, I added myself as a Manager for each Admin Set. Then I could see all the queues.
I thought I was able to see items in the queue without being a Manager — there were even some Grad School Review items in my queue from March tests — since I was part of the Repo Managers group?
From wickr:
> this is from the Samvera docs: "In the Dashboard, click on “Tasks,” then “Review Submissions.” You will see a list of works requiring your review. Repository Admins will see all works in the repository, other reviewers will only see works in Admin Sets for which they are designated Managers."
So:
1. I am not sure what altered the Repo Managers group permissions and what impact it would have on production.
2. It looks like there are new Admin Sets for GTDs and HCTs? I don’t know how that is related to the above but if appropriate, it would be nice if old/unused ones were removed.
3. Whatever the situation, the goal is to have staging and prod match.
|
non_build
|
staging cleanup admin permissions change i deposited into each worktype to test workflows while the blue status “button” listed the correct stage of review grad school review library review pending review honors college review and i could approve advance as needed from the show page none of the items showed up in the review queue for me after discussion with wickr i added myself as a manager for each admin set then i could see all the queues i thought i was able to see items in the queue without being a manager — there were even some grad school review items in my queue from march tests — since i was part of the repo managers group from wickr this is from the samvera docs in the dashboard click on “tasks ” then “review submissions ” you will see a list of works requiring your review repository admins will see all works in the repository other reviewers will only see works in admin sets for which they are designated managers so i am not sure what altered the repo managers group permissions and what impact it would have on production it looks like there are new admin sets for gtds and hcts i don’t know how that is related to the above but if appropriate it would be nice if old unused ones were removed whatever the situation the goal is to have staging and prod match
| 0
|
73,647
| 19,728,488,175
|
IssuesEvent
|
2022-01-13 22:40:49
|
ClangBuiltLinux/linux
|
https://api.github.com/repos/ClangBuiltLinux/linux
|
closed
|
-Wframe-larger-than= in drivers/staging/greybus/audio_topology.c
|
[BUG] linux [ARCH] arm64 [ARCH] x86_64 [PATCH] Accepted -Wframe-larger-than= [CONFIG] allmodconfig CONFIG_WERROR Clean build
|
With `clang-11` and `clang-12`, I see the following error with x86_64 allmodconfig:
```
drivers/staging/greybus/audio_topology.c:977:12: error: stack frame size of 3304 bytes in function 'gbaudio_tplg_create_widget' [-Werror,-Wframe-larger-than=]
static int gbaudio_tplg_create_widget(struct gbaudio_module_info *module,
^
1 error generated.
```
@arndb sent [a patch](https://lore.kernel.org/r/20210103223541.2790855-1-arnd@kernel.org/) for this earlier this year that had some feedback but I never saw a follow up. I addressed the comments, verified the error was fixed, and sent a v2: https://lore.kernel.org/r/20211209195141.1165233-1-nathan@kernel.org/
|
1.0
|
-Wframe-larger-than= in drivers/staging/greybus/audio_topology.c - With `clang-11` and `clang-12`, I see the following error with x86_64 allmodconfig:
```
drivers/staging/greybus/audio_topology.c:977:12: error: stack frame size of 3304 bytes in function 'gbaudio_tplg_create_widget' [-Werror,-Wframe-larger-than=]
static int gbaudio_tplg_create_widget(struct gbaudio_module_info *module,
^
1 error generated.
```
@arndb sent [a patch](https://lore.kernel.org/r/20210103223541.2790855-1-arnd@kernel.org/) for this earlier this year that had some feedback but I never saw a follow up. I addressed the comments, verified the error was fixed, and sent a v2: https://lore.kernel.org/r/20211209195141.1165233-1-nathan@kernel.org/
|
build
|
wframe larger than in drivers staging greybus audio topology c with clang and clang i see the following error with allmodconfig drivers staging greybus audio topology c error stack frame size of bytes in function gbaudio tplg create widget static int gbaudio tplg create widget struct gbaudio module info module error generated arndb sent for this earlier this year that had some feedback but i never saw a follow up i addressed the comments verified the error was fixed and sent a
| 1
|
334,831
| 29,993,177,231
|
IssuesEvent
|
2023-06-26 01:39:09
|
dotnet/machinelearning-modelbuilder
|
https://api.github.com/repos/dotnet/machinelearning-modelbuilder
|
closed
|
Sentence similarity: Project debug failed with error "Two TorchSharp runtime packages have been referenced (both libtorch-cpu and libtorch-cuda)".
|
Priority:0 Reported by: Test
|
**System Information (please complete the following information):**
Windows OS: Windows-11-Enterprise-22H2
ML.Net Model Builder 2022: 17.15.0.2322401 (Main Build)
Microsoft Visual Studio Enterprise: 2022(17.5.2)
.Net: 6.0
**Describe the bug**
- On which step of the process did you run into an issue:
Project running failed with error "Two TorchSharp runtime packages have been referenced (both libtorch-cpu and libtorch-cuda)".
**TestMatrix**
[home-depot-sentence-similarity-small.csv](https://github.com/dotnet/machinelearning-modelbuilder/files/10534387/home-depot-sentence-similarity-small.csv)
**To Reproduce**
Steps to reproduce the behavior:
1. Select Create a new project from the Visual Studio start window.
2. Choose the C# Console App (.NET Core) project template.
3. Add model builder by right click on the project.
4. Select **"Sentence similarity"** scenario.
5. On Data page, choose the attached .csv data source and set the three options like below.

6. On the Train page, click "Start training" button to complete the training.
7. Press "F5" to debug the project.
**Expected behavior**
Project debug successful.
**Screenshot**

**Additional context**
1. There are many errors in "consumption.cs" and "training.cs" files.


2. The issue is not repro in the generated Console App and Web API projects.
|
1.0
|
Sentence similarity: Project debug failed with error "Two TorchSharp runtime packages have been referenced (both libtorch-cpu and libtorch-cuda)". - **System Information (please complete the following information):**
Windows OS: Windows-11-Enterprise-22H2
ML.Net Model Builder 2022: 17.15.0.2322401 (Main Build)
Microsoft Visual Studio Enterprise: 2022(17.5.2)
.Net: 6.0
**Describe the bug**
- On which step of the process did you run into an issue:
Project running failed with error "Two TorchSharp runtime packages have been referenced (both libtorch-cpu and libtorch-cuda)".
**TestMatrix**
[home-depot-sentence-similarity-small.csv](https://github.com/dotnet/machinelearning-modelbuilder/files/10534387/home-depot-sentence-similarity-small.csv)
**To Reproduce**
Steps to reproduce the behavior:
1. Select Create a new project from the Visual Studio start window.
2. Choose the C# Console App (.NET Core) project template.
3. Add model builder by right click on the project.
4. Select **"Sentence similarity"** scenario.
5. On Data page, choose the attached .csv data source and set the three options like below.

6. On the Train page, click "Start training" button to complete the training.
7. Press "F5" to debug the project.
**Expected behavior**
Project debug successful.
**Screenshot**

**Additional context**
1. There are many errors in "consumption.cs" and "training.cs" files.


2. The issue is not repro in the generated Console App and Web API projects.
|
non_build
|
sentence similarity project debug failed with error two torchsharp runtime packages have been referenced both libtorch cpu and libtorch cuda system information please complete the following information windows os windows enterprise ml net model builder main build microsoft visual studio enterprise net describe the bug on which step of the process did you run into an issue project running failed with error two torchsharp runtime packages have been referenced both libtorch cpu and libtorch cuda testmatrix to reproduce steps to reproduce the behavior select create a new project from the visual studio start window choose the c console app net core project template add model builder by right click on the project select sentence similarity scenario on data page choose the attached csv data source and set the three options like below on the train page click start training button to complete the training press to debug the project expected behavior project debug successful screenshot additional context there are many errors in consumption cs and training cs files the issue is not repro in the generated console app and web api projects
| 0
|
748,672
| 26,132,423,600
|
IssuesEvent
|
2022-12-29 07:33:08
|
pystardust/ani-cli
|
https://api.github.com/repos/pystardust/ani-cli
|
closed
|
GogoAnime as host instead of AnimixPlay
|
type: feature request priority 4: wishlist
|

Since AnimixPlay is gone forever,
please update the stdout from:
`Episodes not released yet!`
to
`Animixplay.to is no longer available. Please wait until we change the web scraper to a different domain. We apologize for the inconvenience.`
there's no other choice than gogoanime. 😅
|
1.0
|
GogoAnime as host instead of AnimixPlay - 
Since AnimixPlay is gone forever,
please update the stdout from:
`Episodes not released yet!`
to
`Animixplay.to is no longer available. Please wait until we change the web scraper to a different domain. We apologize for the inconvenience.`
there's no other choice than gogoanime. 😅
|
non_build
|
gogoanime as host instead of animixplay since animixplay is gone forever please update the stdout from episodes not released yet to animixplay to is no longer available please wait until we change the web scraper to a different domain we apologize for the inconvenience there s no other choice than gogoanime 😅
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.