Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
4,464
7,331,139,677
IssuesEvent
2018-03-05 12:28:28
zotero/zotero
https://api.github.com/repos/zotero/zotero
closed
Show "Automatically update citations" in document preferences after document creation
Blocker Word Processor Integration
People should be able to turn on delayed citations manually after document creation. It shouldn't show up during document creation, but it should show up if you return to the doc prefs. Discussed here: https://forums.zotero.org/discussion/comment/302684/#Comment_302684
1.0
Show "Automatically update citations" in document preferences after document creation - People should be able to turn on delayed citations manually after document creation. It shouldn't show up during document creation, but it should show up if you return to the doc prefs. Discussed here: https://forums.zotero.org/discussion/comment/302684/#Comment_302684
process
show automatically update citations in document preferences after document creation people should be able to turn on delayed citations manually after document creation it shouldn t show up during document creation but it should show up if you return to the doc prefs discussed here
1
5,176
7,960,135,641
IssuesEvent
2018-07-13 05:38:11
Rokid/ShadowNode
https://api.github.com/repos/Rokid/ShadowNode
closed
child_process: process.send causes memory leaks
bug child_process
```js // parent.js var child = require('child_process').fork(__dirname + '/child.js', { env: { isSubprocess: 'true', } }) child.on('message', data => { // console.log(data.toString()) }) //child.js setInterval(() => { process.send(Math.random()) }, 0) ``` child's memory increases fast
1.0
child_process: process.send causes memory leaks - ```js // parent.js var child = require('child_process').fork(__dirname + '/child.js', { env: { isSubprocess: 'true', } }) child.on('message', data => { // console.log(data.toString()) }) //child.js setInterval(() => { process.send(Math.random()) }, 0) ``` child's memory increases fast
process
child process process send causes memory leaks js parent js var child require child process fork dirname child js env issubprocess true child on message data console log data tostring child js setinterval process send math random child s memory increases fast
1
10,017
13,043,914,329
IssuesEvent
2020-07-29 03:02:12
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `Uncompress` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `Uncompress` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @sticnarf ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `Uncompress` from TiDB - ## Description Port the scalar function `Uncompress` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @sticnarf ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function uncompress from tidb description port the scalar function uncompress from tidb to coprocessor score mentor s sticnarf recommended skills rust programming learning materials already implemented expressions ported from tidb
1
16,836
9,536,669,789
IssuesEvent
2019-04-30 10:20:25
Garados007/Werwolf
https://api.github.com/repos/Garados007/Werwolf
closed
Optimiere Abrufe bei Spielrundenwechsel
difficult performance
Bei einem Wechsel der aktuellen Spielrunde werden ein Großteil der Daten (z.B. Chateinträge) verworfen und müssen neu abgerufen werden. Ein Teil davon ändert sich aber nicht in der nächsten Runde und soll nur ausgeblendet werden oder ungültige registrierte periodische Abrfragen existieren. Diese Abfragen lassen sich optimieren.
True
Optimiere Abrufe bei Spielrundenwechsel - Bei einem Wechsel der aktuellen Spielrunde werden ein Großteil der Daten (z.B. Chateinträge) verworfen und müssen neu abgerufen werden. Ein Teil davon ändert sich aber nicht in der nächsten Runde und soll nur ausgeblendet werden oder ungültige registrierte periodische Abrfragen existieren. Diese Abfragen lassen sich optimieren.
non_process
optimiere abrufe bei spielrundenwechsel bei einem wechsel der aktuellen spielrunde werden ein großteil der daten z b chateinträge verworfen und müssen neu abgerufen werden ein teil davon ändert sich aber nicht in der nächsten runde und soll nur ausgeblendet werden oder ungültige registrierte periodische abrfragen existieren diese abfragen lassen sich optimieren
0
54,772
13,920,332,315
IssuesEvent
2020-10-21 10:17:05
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Cannot convert from UUID to JSON in H2's JSON_OBJECT() and related functions
C: DB: H2 C: Functionality E: All Editions P: Medium R: Fixed T: Defect
### Expected behavior Query result containing records with UUID successfully mapped to DTO via JSON document. ### Actual behavior Exception is thrown when JOOQ attempts to map the record: ``` org.springframework.dao.DataIntegrityViolationException: jOOQ; SQL [select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME"]; Data conversion error converting "UUID to JSON" Data conversion error converting "UUID to JSON"; SQL statement: select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]; nested exception is org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON" Data conversion error converting "UUID to JSON"; SQL statement: select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200] at org.jooq_3.14.0.DEFAULT.debug(Unknown Source) at org.springframework.jdbc.support.SQLStateSQLExceptionTranslator.doTranslate(SQLStateSQLExceptionTranslator.java:104) at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:72) at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.translate(JooqExceptionTranslator.java:92) at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.handle(JooqExceptionTranslator.java:81) at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.exception(JooqExceptionTranslator.java:55) at org.jooq.impl.ExecuteListeners.exception(ExecuteListeners.java:274) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:390) at org.jooq.impl.AbstractResultQuery.fetch(AbstractResultQuery.java:333) at org.jooq.impl.AbstractResultQuery.fetchInto(AbstractResultQuery.java:1545) at org.jooq.impl.SelectImpl.fetchInto(SelectImpl.java:3936) at com.example.uuidjson.JooqParentRepository.list(JooqParentRepository.java:38) at com.example.uuidjson.JooqParentRepository$$FastClassBySpringCGLIB$$93c0a175.invoke(<generated>) at org.springframework.cglib.proxy.MethodProxy.invoke(MethodProxy.java:218) at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.invokeJoinpoint(CglibAopProxy.java:771) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:163) at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749) at org.springframework.transaction.interceptor.TransactionAspectSupport.invokeWithinTransaction(TransactionAspectSupport.java:367) at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:118) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:186) at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749) at org.springframework.aop.framework.CglibAopProxy$DynamicAdvisedInterceptor.intercept(CglibAopProxy.java:691) at com.example.uuidjson.JooqParentRepository$$EnhancerBySpringCGLIB$$92b1c668.list(<generated>) at com.example.uuidjson.JooqParentRepositoryTest.parentWithChildrenFetched(JooqParentRepositoryTest.java:25) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:564) at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:686) at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131) at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84) at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115) at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$6(TestMethodTestDescriptor.java:212) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:208) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:137) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:71) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:135) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80) at java.base/java.util.ArrayList.forEach(ArrayList.java:1510) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80) at java.base/java.util.ArrayList.forEach(ArrayList.java:1510) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:32) at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57) at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:51) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:248) at org.junit.platform.launcher.core.DefaultLauncher.lambda$execute$5(DefaultLauncher.java:211) at org.junit.platform.launcher.core.DefaultLauncher.withInterceptedStreams(DefaultLauncher.java:226) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:199) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:132) at com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:71) at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33) at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220) at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53) Caused by: org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON" Data conversion error converting "UUID to JSON"; SQL statement: select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200] at org.h2.message.DbException.getJdbcSQLException(DbException.java:457) at org.h2.message.DbException.getJdbcSQLException(DbException.java:429) at org.h2.message.DbException.get(DbException.java:205) at org.h2.message.DbException.get(DbException.java:181) at org.h2.value.Value.getDataConversionError(Value.java:1504) at org.h2.value.Value.convertToJson(Value.java:1439) at org.h2.value.Value.convertTo(Value.java:861) at org.h2.value.Value.convertTo(Value.java:737) at org.h2.expression.function.Function.jsonObjectAppend(Function.java:2368) at org.h2.expression.function.Function.jsonObject(Function.java:2351) at org.h2.expression.function.Function.getSimpleValue(Function.java:1153) at org.h2.expression.function.Function.getValueWithArgs(Function.java:1274) at org.h2.expression.function.Function.getValue(Function.java:672) at org.h2.expression.aggregate.Aggregate.updateAggregate(Aggregate.java:205) at org.h2.expression.aggregate.AbstractAggregate.updateAggregate(AbstractAggregate.java:280) at org.h2.expression.analysis.DataAnalysisOperation.updateAggregate(DataAnalysisOperation.java:252) at org.h2.expression.function.Function.updateAggregate(Function.java:3049) at org.h2.command.dml.Select.updateAgg(Select.java:545) at org.h2.command.dml.Select.gatherGroup(Select.java:525) at org.h2.command.dml.Select.queryGroup(Select.java:487) at org.h2.command.dml.Select.queryWithoutCache(Select.java:839) at org.h2.command.dml.Query.queryWithoutCacheLazyCheck(Query.java:201) at org.h2.command.dml.Query.query(Query.java:489) at org.h2.command.dml.Query.query(Query.java:451) at org.h2.command.CommandContainer.query(CommandContainer.java:285) at org.h2.command.Command.executeQuery(Command.java:195) at org.h2.jdbc.JdbcPreparedStatement.execute(JdbcPreparedStatement.java:233) at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:214) at org.jooq.impl.Tools.executeStatementAndGetFirstResultSet(Tools.java:4186) at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:279) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:375) ... 79 more ``` ### Steps to reproduce the problem 1. Create table with UUID as primary key: ``` create table parent( id UUID constraint pk_parent_id primary key, name varchar(100) ); ``` 2. Run code generation, check that generated field is of UUID type: ``` public final TableField<ParentRecord, UUID> ID = createField(DSL.name("ID"), SQLDataType.UUID.nullable(false), this, ""); ``` 3. Create target DTO for record mapper: ``` @Getter @RequiredArgsConstructor public class Parent { private final UUID id; private final String name; private final Collection<Child> children; } ``` 4. Run a query mapping its result into DTO using JSON as described here: https://blog.jooq.org/2020/10/09/nesting-collections-with-jooq-3-14s-sql-xml-or-sql-json-support/ ``` dsl.select( jsonObject( jsonEntry("id", PARENT.ID), jsonEntry("name", PARENT.NAME), jsonEntry("children", jsonArrayAgg( jsonObject( jsonEntry("parentId", CHILD.PARENT_ID), jsonEntry("name", CHILD.NAME) ) ).orderBy(PARENT.NAME)) ) ) .from(PARENT) .leftJoin(CHILD).on(CHILD.PARENT_ID.eq(PARENT.ID)) .groupBy(PARENT.ID, PARENT.NAME) .orderBy(PARENT.ID, PARENT.NAME) .fetchInto(Parent.class); ``` ### Versions - jOOQ: 3.14.0 - Java: 11 - Database: h2 - OS: Windows - JDBC Driver: org.h2.Driver
1.0
Cannot convert from UUID to JSON in H2's JSON_OBJECT() and related functions - ### Expected behavior Query result containing records with UUID successfully mapped to DTO via JSON document. ### Actual behavior Exception is thrown when JOOQ attempts to map the record: ``` org.springframework.dao.DataIntegrityViolationException: jOOQ; SQL [select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME"]; Data conversion error converting "UUID to JSON" Data conversion error converting "UUID to JSON"; SQL statement: select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]; nested exception is org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON" Data conversion error converting "UUID to JSON"; SQL statement: select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200] at org.jooq_3.14.0.DEFAULT.debug(Unknown Source) at org.springframework.jdbc.support.SQLStateSQLExceptionTranslator.doTranslate(SQLStateSQLExceptionTranslator.java:104) at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:72) at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.translate(JooqExceptionTranslator.java:92) at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.handle(JooqExceptionTranslator.java:81) at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.exception(JooqExceptionTranslator.java:55) at org.jooq.impl.ExecuteListeners.exception(ExecuteListeners.java:274) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:390) at org.jooq.impl.AbstractResultQuery.fetch(AbstractResultQuery.java:333) at org.jooq.impl.AbstractResultQuery.fetchInto(AbstractResultQuery.java:1545) at org.jooq.impl.SelectImpl.fetchInto(SelectImpl.java:3936) at com.example.uuidjson.JooqParentRepository.list(JooqParentRepository.java:38) at com.example.uuidjson.JooqParentRepository$$FastClassBySpringCGLIB$$93c0a175.invoke(<generated>) at org.springframework.cglib.proxy.MethodProxy.invoke(MethodProxy.java:218) at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.invokeJoinpoint(CglibAopProxy.java:771) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:163) at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749) at org.springframework.transaction.interceptor.TransactionAspectSupport.invokeWithinTransaction(TransactionAspectSupport.java:367) at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:118) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:186) at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749) at org.springframework.aop.framework.CglibAopProxy$DynamicAdvisedInterceptor.intercept(CglibAopProxy.java:691) at com.example.uuidjson.JooqParentRepository$$EnhancerBySpringCGLIB$$92b1c668.list(<generated>) at com.example.uuidjson.JooqParentRepositoryTest.parentWithChildrenFetched(JooqParentRepositoryTest.java:25) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:564) at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:686) at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131) at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84) at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115) at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$6(TestMethodTestDescriptor.java:212) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:208) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:137) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:71) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:135) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80) at java.base/java.util.ArrayList.forEach(ArrayList.java:1510) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80) at java.base/java.util.ArrayList.forEach(ArrayList.java:1510) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:32) at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57) at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:51) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:248) at org.junit.platform.launcher.core.DefaultLauncher.lambda$execute$5(DefaultLauncher.java:211) at org.junit.platform.launcher.core.DefaultLauncher.withInterceptedStreams(DefaultLauncher.java:226) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:199) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:132) at com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:71) at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33) at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220) at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53) Caused by: org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON" Data conversion error converting "UUID to JSON"; SQL statement: select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200] at org.h2.message.DbException.getJdbcSQLException(DbException.java:457) at org.h2.message.DbException.getJdbcSQLException(DbException.java:429) at org.h2.message.DbException.get(DbException.java:205) at org.h2.message.DbException.get(DbException.java:181) at org.h2.value.Value.getDataConversionError(Value.java:1504) at org.h2.value.Value.convertToJson(Value.java:1439) at org.h2.value.Value.convertTo(Value.java:861) at org.h2.value.Value.convertTo(Value.java:737) at org.h2.expression.function.Function.jsonObjectAppend(Function.java:2368) at org.h2.expression.function.Function.jsonObject(Function.java:2351) at org.h2.expression.function.Function.getSimpleValue(Function.java:1153) at org.h2.expression.function.Function.getValueWithArgs(Function.java:1274) at org.h2.expression.function.Function.getValue(Function.java:672) at org.h2.expression.aggregate.Aggregate.updateAggregate(Aggregate.java:205) at org.h2.expression.aggregate.AbstractAggregate.updateAggregate(AbstractAggregate.java:280) at org.h2.expression.analysis.DataAnalysisOperation.updateAggregate(DataAnalysisOperation.java:252) at org.h2.expression.function.Function.updateAggregate(Function.java:3049) at org.h2.command.dml.Select.updateAgg(Select.java:545) at org.h2.command.dml.Select.gatherGroup(Select.java:525) at org.h2.command.dml.Select.queryGroup(Select.java:487) at org.h2.command.dml.Select.queryWithoutCache(Select.java:839) at org.h2.command.dml.Query.queryWithoutCacheLazyCheck(Query.java:201) at org.h2.command.dml.Query.query(Query.java:489) at org.h2.command.dml.Query.query(Query.java:451) at org.h2.command.CommandContainer.query(CommandContainer.java:285) at org.h2.command.Command.executeQuery(Command.java:195) at org.h2.jdbc.JdbcPreparedStatement.execute(JdbcPreparedStatement.java:233) at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:214) at org.jooq.impl.Tools.executeStatementAndGetFirstResultSet(Tools.java:4186) at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:279) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:375) ... 79 more ``` ### Steps to reproduce the problem 1. Create table with UUID as primary key: ``` create table parent( id UUID constraint pk_parent_id primary key, name varchar(100) ); ``` 2. Run code generation, check that generated field is of UUID type: ``` public final TableField<ParentRecord, UUID> ID = createField(DSL.name("ID"), SQLDataType.UUID.nullable(false), this, ""); ``` 3. Create target DTO for record mapper: ``` @Getter @RequiredArgsConstructor public class Parent { private final UUID id; private final String name; private final Collection<Child> children; } ``` 4. Run a query mapping its result into DTO using JSON as described here: https://blog.jooq.org/2020/10/09/nesting-collections-with-jooq-3-14s-sql-xml-or-sql-json-support/ ``` dsl.select( jsonObject( jsonEntry("id", PARENT.ID), jsonEntry("name", PARENT.NAME), jsonEntry("children", jsonArrayAgg( jsonObject( jsonEntry("parentId", CHILD.PARENT_ID), jsonEntry("name", CHILD.NAME) ) ).orderBy(PARENT.NAME)) ) ) .from(PARENT) .leftJoin(CHILD).on(CHILD.PARENT_ID.eq(PARENT.ID)) .groupBy(PARENT.ID, PARENT.NAME) .orderBy(PARENT.ID, PARENT.NAME) .fetchInto(Parent.class); ``` ### Versions - jOOQ: 3.14.0 - Java: 11 - Database: h2 - OS: Windows - JDBC Driver: org.h2.Driver
non_process
cannot convert from uuid to json in s json object and related functions expected behavior query result containing records with uuid successfully mapped to dto via json document actual behavior exception is thrown when jooq attempts to map the record org springframework dao dataintegrityviolationexception jooq sql data conversion error converting uuid to json data conversion error converting uuid to json sql statement select json object key value parent id key value parent name key value json arrayagg json object key value child parent id key value child name order by parent name from parent left outer join child on child parent id parent id group by parent id parent name order by parent id parent name nested exception is org jdbc jdbcsqldataexception data conversion error converting uuid to json data conversion error converting uuid to json sql statement select json object key value parent id key value parent name key value json arrayagg json object key value child parent id key value child name order by parent name from parent left outer join child on child parent id parent id group by parent id parent name order by parent id parent name at org jooq default debug unknown source at org springframework jdbc support sqlstatesqlexceptiontranslator dotranslate sqlstatesqlexceptiontranslator java at org springframework jdbc support abstractfallbacksqlexceptiontranslator translate abstractfallbacksqlexceptiontranslator java at org springframework boot autoconfigure jooq jooqexceptiontranslator translate jooqexceptiontranslator java at org springframework boot autoconfigure jooq jooqexceptiontranslator handle jooqexceptiontranslator java at org springframework boot autoconfigure jooq jooqexceptiontranslator exception jooqexceptiontranslator java at org jooq impl executelisteners exception executelisteners java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractresultquery fetch abstractresultquery java at org jooq impl abstractresultquery fetchinto abstractresultquery java at org jooq impl selectimpl fetchinto selectimpl java at com example uuidjson jooqparentrepository list jooqparentrepository java at com example uuidjson jooqparentrepository fastclassbyspringcglib invoke at org springframework cglib proxy methodproxy invoke methodproxy java at org springframework aop framework cglibaopproxy cglibmethodinvocation invokejoinpoint cglibaopproxy java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop framework cglibaopproxy cglibmethodinvocation proceed cglibaopproxy java at org springframework transaction interceptor transactionaspectsupport invokewithintransaction transactionaspectsupport java at org springframework transaction interceptor transactioninterceptor invoke transactioninterceptor java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop framework cglibaopproxy cglibmethodinvocation proceed cglibaopproxy java at org springframework aop framework cglibaopproxy dynamicadvisedinterceptor intercept cglibaopproxy java at com example uuidjson jooqparentrepository enhancerbyspringcglib list at com example uuidjson jooqparentrepositorytest parentwithchildrenfetched jooqparentrepositorytest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit platform commons util reflectionutils invokemethod reflectionutils java at org junit jupiter engine execution methodinvocation proceed methodinvocation java at org junit jupiter engine execution invocationinterceptorchain validatinginvocation proceed invocationinterceptorchain java at org junit jupiter engine extension timeoutextension intercept timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestablemethod timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestmethod timeoutextension java at org junit jupiter engine execution executableinvoker reflectiveinterceptorcall lambda ofvoidmethod executableinvoker java at org junit jupiter engine execution executableinvoker lambda invoke executableinvoker java at org junit jupiter engine execution invocationinterceptorchain interceptedinvocation proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain chainandinvoke invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain invoke invocationinterceptorchain java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine descriptor testmethodtestdescriptor lambda invoketestmethod testmethodtestdescriptor java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit jupiter engine descriptor testmethodtestdescriptor invoketestmethod testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at java base java util arraylist foreach arraylist java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice invokeall samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at java base java util arraylist foreach arraylist java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice invokeall samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice submit samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical hierarchicaltestexecutor execute hierarchicaltestexecutor java at org junit platform engine support hierarchical hierarchicaltestengine execute hierarchicaltestengine java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher lambda execute defaultlauncher java at org junit platform launcher core defaultlauncher withinterceptedstreams defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at com intellij startrunnerwithargs java at com intellij rt junit ideatestrunner repeater startrunnerwithargs ideatestrunner java at com intellij rt junit junitstarter preparestreamsandstart junitstarter java at com intellij rt junit junitstarter main junitstarter java caused by org jdbc jdbcsqldataexception data conversion error converting uuid to json data conversion error converting uuid to json sql statement select json object key value parent id key value parent name key value json arrayagg json object key value child parent id key value child name order by parent name from parent left outer join child on child parent id parent id group by parent id parent name order by parent id parent name at org message dbexception getjdbcsqlexception dbexception java at org message dbexception getjdbcsqlexception dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org value value getdataconversionerror value java at org value value converttojson value java at org value value convertto value java at org value value convertto value java at org expression function function jsonobjectappend function java at org expression function function jsonobject function java at org expression function function getsimplevalue function java at org expression function function getvaluewithargs function java at org expression function function getvalue function java at org expression aggregate aggregate updateaggregate aggregate java at org expression aggregate abstractaggregate updateaggregate abstractaggregate java at org expression analysis dataanalysisoperation updateaggregate dataanalysisoperation java at org expression function function updateaggregate function java at org command dml select updateagg select java at org command dml select gathergroup select java at org command dml select querygroup select java at org command dml select querywithoutcache select java at org command dml query querywithoutcachelazycheck query java at org command dml query query query java at org command dml query query query java at org command commandcontainer query commandcontainer java at org command command executequery command java at org jdbc jdbcpreparedstatement execute jdbcpreparedstatement java at org jooq tools jdbc defaultpreparedstatement execute defaultpreparedstatement java at org jooq impl tools executestatementandgetfirstresultset tools java at org jooq impl abstractresultquery execute abstractresultquery java at org jooq impl abstractquery execute abstractquery java more steps to reproduce the problem create table with uuid as primary key create table parent id uuid constraint pk parent id primary key name varchar run code generation check that generated field is of uuid type public final tablefield id createfield dsl name id sqldatatype uuid nullable false this create target dto for record mapper getter requiredargsconstructor public class parent private final uuid id private final string name private final collection children run a query mapping its result into dto using json as described here dsl select jsonobject jsonentry id parent id jsonentry name parent name jsonentry children jsonarrayagg jsonobject jsonentry parentid child parent id jsonentry name child name orderby parent name from parent leftjoin child on child parent id eq parent id groupby parent id parent name orderby parent id parent name fetchinto parent class versions jooq java database os windows jdbc driver org driver
0
17,812
23,739,991,520
IssuesEvent
2022-08-31 11:34:59
Tencent/tdesign-miniprogram
https://api.github.com/repos/Tencent/tdesign-miniprogram
closed
[dialog] 文档或示例中缺少open-type-event的调用案例
good first issue in process
### 这个功能解决了什么问题 dialog的确认键如果想使用button的开放能力获取手机号,用哪个字段、如何声明getPhoneNumber呢 ### 你建议的方案是什么 示例中给个使用的案例吧
1.0
[dialog] 文档或示例中缺少open-type-event的调用案例 - ### 这个功能解决了什么问题 dialog的确认键如果想使用button的开放能力获取手机号,用哪个字段、如何声明getPhoneNumber呢 ### 你建议的方案是什么 示例中给个使用的案例吧
process
文档或示例中缺少open type event的调用案例 这个功能解决了什么问题 dialog的确认键如果想使用button的开放能力获取手机号,用哪个字段、如何声明getphonenumber呢 你建议的方案是什么 示例中给个使用的案例吧
1
105,725
9,100,269,255
IssuesEvent
2019-02-20 08:01:42
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
opened
Test : ApiV1OrgsIdPutOrgorgplanbasicUsercDisallowAbact2
test
Project : Test Job : Default Env : Default Category : null Tags : null Severity : null Region : US_WEST Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/ Request : { "billingEmail" : "katrina.reynolds@yahoo.com", "company" : "Kuvalis-Kuvalis", "createdBy" : "", "createdDate" : "", "description" : "cXD8m8jM", "id" : "", "inactive" : false, "location" : "cXD8m8jM", "modifiedBy" : "", "modifiedDate" : "", "name" : "cXD8m8jM", "orgPlan" : "TEAM", "orgType" : "TEAM", "version" : "" } Response : I/O error on PUT request for "http://13.56.210.25/api/v1/api/v1/orgs/": Read timed out; nested exception is java.net.SocketTimeoutException: Read timed out Logs : Assertion [@StatusCode == 401 OR @StatusCode == 403 OR @Response.errors == true] resolved-to [500 == 401 OR 500 == 403 OR == true] result [Failed] --- FX Bot ---
1.0
Test : ApiV1OrgsIdPutOrgorgplanbasicUsercDisallowAbact2 - Project : Test Job : Default Env : Default Category : null Tags : null Severity : null Region : US_WEST Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/ Request : { "billingEmail" : "katrina.reynolds@yahoo.com", "company" : "Kuvalis-Kuvalis", "createdBy" : "", "createdDate" : "", "description" : "cXD8m8jM", "id" : "", "inactive" : false, "location" : "cXD8m8jM", "modifiedBy" : "", "modifiedDate" : "", "name" : "cXD8m8jM", "orgPlan" : "TEAM", "orgType" : "TEAM", "version" : "" } Response : I/O error on PUT request for "http://13.56.210.25/api/v1/api/v1/orgs/": Read timed out; nested exception is java.net.SocketTimeoutException: Read timed out Logs : Assertion [@StatusCode == 401 OR @StatusCode == 403 OR @Response.errors == true] resolved-to [500 == 401 OR 500 == 403 OR == true] result [Failed] --- FX Bot ---
non_process
test project test job default env default category null tags null severity null region us west result fail status code headers endpoint request billingemail katrina reynolds yahoo com company kuvalis kuvalis createdby createddate description id inactive false location modifiedby modifieddate name orgplan team orgtype team version response i o error on put request for read timed out nested exception is java net sockettimeoutexception read timed out logs assertion resolved to result fx bot
0
77,675
3,507,208,456
IssuesEvent
2016-01-08 11:54:51
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Crash Alert (BB #655)
Category: Crash migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** PadreWoW **Original Date:** 20.08.2014 09:22:23 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** closed **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/655 <hr> 2014-08-20 13:20:33 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:39 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:44 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:49 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:57 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:21:05 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
1.0
Crash Alert (BB #655) - This issue was migrated from bitbucket. **Original Reporter:** PadreWoW **Original Date:** 20.08.2014 09:22:23 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** closed **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/655 <hr> 2014-08-20 13:20:33 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:39 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:44 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:49 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:20:57 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)! 2014-08-20 13:21:05 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
non_process
crash alert bb this issue was migrated from bitbucket original reporter padrewow original date gmt original priority major original type bug original state closed direct link crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type
0
453,359
13,068,578,819
IssuesEvent
2020-07-31 04:01:52
ProjectSidewalk/SidewalkWebpage
https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage
closed
In Version 2.1, No Smooth Animated Interpolation Between Locations When Double Clicking
Priority: Low wontfix
In version 2.1, there is no smooth animated interpolation between locations when double clicking to move. See below. Is there some function parameter we have to pass to enable this smooth movement animation on the double click to move? It's enabled when you click the arrow to move. ![image](https://cloud.githubusercontent.com/assets/1621749/24199024/89151f4e-0ede-11e7-9e59-e78c0e913966.png)
1.0
In Version 2.1, No Smooth Animated Interpolation Between Locations When Double Clicking - In version 2.1, there is no smooth animated interpolation between locations when double clicking to move. See below. Is there some function parameter we have to pass to enable this smooth movement animation on the double click to move? It's enabled when you click the arrow to move. ![image](https://cloud.githubusercontent.com/assets/1621749/24199024/89151f4e-0ede-11e7-9e59-e78c0e913966.png)
non_process
in version no smooth animated interpolation between locations when double clicking in version there is no smooth animated interpolation between locations when double clicking to move see below is there some function parameter we have to pass to enable this smooth movement animation on the double click to move it s enabled when you click the arrow to move
0
4,638
7,482,326,321
IssuesEvent
2018-04-05 00:39:08
UnbFeelings/unb-feelings-GQA
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
closed
Mudar o template da descrição de atividades
process wiki
Mudar o template atual que é o seguinte: <table> <tbody> <tr> <td><b>Descrição</b></td> <td>Texto</td> </tr> <tr> <td><b>Papéis</b></td> <td>Texto</td> </tr> <tr> <td><b>Entradas</b></td> <td>Texto</td> </tr> <tr> <td><b>Saídas</b></td> <td>Texto</td> </tr> </tbody> </table> Para: <table> <tbody> <tr> <td><b>Descrição</b></td> <td>Texto</td> </tr> <tr> <td><b>Responsáveis</b></td> <td>Texto</td> </tr> <tr> <td><b>Envolvidos</b></td> <td>Texto</td> </tr> <tr> <td><b>Entradas</b></td> <td>Texto</td> </tr> <tr> <td><b>Saídas</b></td> <td>Texto</td> </tr> </tbody> </table> Resumindo: Na descrição das atividades,colocar : - Responsáveis - Envolvidos Em substituição aos papéis
1.0
Mudar o template da descrição de atividades - Mudar o template atual que é o seguinte: <table> <tbody> <tr> <td><b>Descrição</b></td> <td>Texto</td> </tr> <tr> <td><b>Papéis</b></td> <td>Texto</td> </tr> <tr> <td><b>Entradas</b></td> <td>Texto</td> </tr> <tr> <td><b>Saídas</b></td> <td>Texto</td> </tr> </tbody> </table> Para: <table> <tbody> <tr> <td><b>Descrição</b></td> <td>Texto</td> </tr> <tr> <td><b>Responsáveis</b></td> <td>Texto</td> </tr> <tr> <td><b>Envolvidos</b></td> <td>Texto</td> </tr> <tr> <td><b>Entradas</b></td> <td>Texto</td> </tr> <tr> <td><b>Saídas</b></td> <td>Texto</td> </tr> </tbody> </table> Resumindo: Na descrição das atividades,colocar : - Responsáveis - Envolvidos Em substituição aos papéis
process
mudar o template da descrição de atividades mudar o template atual que é o seguinte descrição texto papéis texto entradas texto saídas texto para descrição texto responsáveis texto envolvidos texto entradas texto saídas texto resumindo na descrição das atividades colocar responsáveis envolvidos em substituição aos papéis
1
137,153
12,746,763,540
IssuesEvent
2020-06-26 16:32:57
Ameelio/letters-api
https://api.github.com/repos/Ameelio/letters-api
closed
Add example Request and example Response bodies to User endpoints
documentation
The following endpoints need example Request and/or Response bodies in the API Documentation Wiki page. GET /api/user GET /api/user/{id} GET /api/users/contacts GET /api/users/letters GET /api/users/org
1.0
Add example Request and example Response bodies to User endpoints - The following endpoints need example Request and/or Response bodies in the API Documentation Wiki page. GET /api/user GET /api/user/{id} GET /api/users/contacts GET /api/users/letters GET /api/users/org
non_process
add example request and example response bodies to user endpoints the following endpoints need example request and or response bodies in the api documentation wiki page get api user get api user id get api users contacts get api users letters get api users org
0
22,229
30,780,346,815
IssuesEvent
2023-07-31 09:34:44
EBIvariation/eva-opentargets
https://api.github.com/repos/EBIvariation/eva-opentargets
opened
Evidence string generation for 2023.09 release
Processing
**Deadline for submission: 8 August 2023** Refer to [documentation](https://github.com/EBIvariation/eva-opentargets/blob/master/docs/generate-evidence-strings.md) for full description of steps.
1.0
Evidence string generation for 2023.09 release - **Deadline for submission: 8 August 2023** Refer to [documentation](https://github.com/EBIvariation/eva-opentargets/blob/master/docs/generate-evidence-strings.md) for full description of steps.
process
evidence string generation for release deadline for submission august refer to for full description of steps
1
770,751
27,054,804,634
IssuesEvent
2023-02-13 15:29:58
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.zeit.de - site is not usable
priority-important browser-firefox-tablet engine-gecko
<!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 4.2.2; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/118150 --> **URL**: https://www.zeit.de/zustimmung?url=https%3A%2F%2Fwww.zeit.de%2Fentdecken%2F2023-02%2Fjudith-holofernes-wochenende-tipps-podcast **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 4.2.2 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Missing items **Steps to Reproduce**: I can't choice the button 'Einverstanden und weiter mit Werbung' <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/2/20e0bea9-62d6-421a-9d17-6532d11934ff.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827153727</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2023/2/bc3bc6fc-b10b-4d72-a83a-edd635629f05) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.zeit.de - site is not usable - <!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 4.2.2; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/118150 --> **URL**: https://www.zeit.de/zustimmung?url=https%3A%2F%2Fwww.zeit.de%2Fentdecken%2F2023-02%2Fjudith-holofernes-wochenende-tipps-podcast **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 4.2.2 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Missing items **Steps to Reproduce**: I can't choice the button 'Einverstanden und weiter mit Werbung' <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/2/20e0bea9-62d6-421a-9d17-6532d11934ff.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827153727</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2023/2/bc3bc6fc-b10b-4d72-a83a-edd635629f05) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
site is not usable url browser version firefox mobile tablet operating system android tested another browser yes chrome problem type site is not usable description missing items steps to reproduce i can t choice the button einverstanden und weiter mit werbung view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
780,618
27,401,754,735
IssuesEvent
2023-03-01 01:34:29
hotosm/fmtm
https://api.github.com/repos/hotosm/fmtm
closed
Add ODK->OSM conversion
enhancement Priority: Must have in progress
Use the odkconvert project conversion program that takes an CSV submission files and converts it to good OSM XML.
1.0
Add ODK->OSM conversion - Use the odkconvert project conversion program that takes an CSV submission files and converts it to good OSM XML.
non_process
add odk osm conversion use the odkconvert project conversion program that takes an csv submission files and converts it to good osm xml
0
20,880
31,466,909,721
IssuesEvent
2023-08-30 03:16:35
BG3-Community-Library-Team/BG3-Community-Library
https://api.github.com/repos/BG3-Community-Library-Team/BG3-Community-Library
closed
Bladesinger Subclass Compatibility Support
Subclass Compatibilty Framework
**Is your feature request related to a problem? Please describe.** We now have permission to add support for the [Bladesinger](https://www.nexusmods.com/baldursgate3/mods/279) subclass. **Describe the solution you'd like** It will need to be implemented in the same way as [PR 39](https://github.com/BG3-Community-Library-Team/BG3-Community-Library/pull/39), which handled some fighter subclasses.
True
Bladesinger Subclass Compatibility Support - **Is your feature request related to a problem? Please describe.** We now have permission to add support for the [Bladesinger](https://www.nexusmods.com/baldursgate3/mods/279) subclass. **Describe the solution you'd like** It will need to be implemented in the same way as [PR 39](https://github.com/BG3-Community-Library-Team/BG3-Community-Library/pull/39), which handled some fighter subclasses.
non_process
bladesinger subclass compatibility support is your feature request related to a problem please describe we now have permission to add support for the subclass describe the solution you d like it will need to be implemented in the same way as which handled some fighter subclasses
0
9,794
12,807,791,993
IssuesEvent
2020-07-03 12:17:12
solid/process
https://api.github.com/repos/solid/process
reopened
Improving the effectiveness of panels
process issue
We [originally introduced](https://github.com/solid/process/pull/24) panels as part of the Solid process ten months ago. Our aim was to drive focused and thoughtful collaboration around specific topics, leading to meaningful contributions to the Solid Specification and ecosystem. We held off on adding a lot of structure around how panels are formed, or how they operate, until we had some real experiences _in the wild_ to learn from. Work towards a normative [Solid Specification](https://github.com/solid/specification) has picked up considerably. We just merged the first [candidate proposal](https://github.com/solid/specification/pull/156) from a panel into the editors draft. We have some real experiences to draw from now, and a lot of work left to be done. It's time to look at how panels have operated to date, and consider any opportunities to improve them. How can we help them drive positive, constructive contributions to the Solid specification and ecosystem in a timely manner? Here are a few considerations and thoughts based on personal observations and discussions with panelists, editors, and others: - There are too many panels, and only a few of them are active and/or productive. We should have less panels with greater coordination and more focus. - There's no clear or consistent organizing structure in active panels. Who is responsible to set expectations, hold or be held accountable, and keep focus where it needs to be? - Active panels aren't focusing enough on the most pressing needs of the specification, or the most important use cases in the real-world. - The editorial team needs to provide more active support to panels, and better communication around specification priorities. _This list is not meant to be exhaustive, only to contribute to the discussion in this thread._ Whatever your role (panelist, community member, editor, etc) - please chime in and share your thoughts on how you'd like to see panels evolve.
1.0
Improving the effectiveness of panels - We [originally introduced](https://github.com/solid/process/pull/24) panels as part of the Solid process ten months ago. Our aim was to drive focused and thoughtful collaboration around specific topics, leading to meaningful contributions to the Solid Specification and ecosystem. We held off on adding a lot of structure around how panels are formed, or how they operate, until we had some real experiences _in the wild_ to learn from. Work towards a normative [Solid Specification](https://github.com/solid/specification) has picked up considerably. We just merged the first [candidate proposal](https://github.com/solid/specification/pull/156) from a panel into the editors draft. We have some real experiences to draw from now, and a lot of work left to be done. It's time to look at how panels have operated to date, and consider any opportunities to improve them. How can we help them drive positive, constructive contributions to the Solid specification and ecosystem in a timely manner? Here are a few considerations and thoughts based on personal observations and discussions with panelists, editors, and others: - There are too many panels, and only a few of them are active and/or productive. We should have less panels with greater coordination and more focus. - There's no clear or consistent organizing structure in active panels. Who is responsible to set expectations, hold or be held accountable, and keep focus where it needs to be? - Active panels aren't focusing enough on the most pressing needs of the specification, or the most important use cases in the real-world. - The editorial team needs to provide more active support to panels, and better communication around specification priorities. _This list is not meant to be exhaustive, only to contribute to the discussion in this thread._ Whatever your role (panelist, community member, editor, etc) - please chime in and share your thoughts on how you'd like to see panels evolve.
process
improving the effectiveness of panels we panels as part of the solid process ten months ago our aim was to drive focused and thoughtful collaboration around specific topics leading to meaningful contributions to the solid specification and ecosystem we held off on adding a lot of structure around how panels are formed or how they operate until we had some real experiences in the wild to learn from work towards a normative has picked up considerably we just merged the first from a panel into the editors draft we have some real experiences to draw from now and a lot of work left to be done it s time to look at how panels have operated to date and consider any opportunities to improve them how can we help them drive positive constructive contributions to the solid specification and ecosystem in a timely manner here are a few considerations and thoughts based on personal observations and discussions with panelists editors and others there are too many panels and only a few of them are active and or productive we should have less panels with greater coordination and more focus there s no clear or consistent organizing structure in active panels who is responsible to set expectations hold or be held accountable and keep focus where it needs to be active panels aren t focusing enough on the most pressing needs of the specification or the most important use cases in the real world the editorial team needs to provide more active support to panels and better communication around specification priorities this list is not meant to be exhaustive only to contribute to the discussion in this thread whatever your role panelist community member editor etc please chime in and share your thoughts on how you d like to see panels evolve
1
460,844
13,219,005,217
IssuesEvent
2020-08-17 09:42:58
MyDataTaiwan/logboard
https://api.github.com/repos/MyDataTaiwan/logboard
opened
When user switch data window by "Today", "This Week" .. buttons, the calendar on top should also sync
QA priority-high
**Steps to Reproduce** 1. User upload data to LogBoard 2. User switch the data window by the buttons "Today", "This Week"...... etc. **Results** ![image](https://user-images.githubusercontent.com/12252921/90381993-c27c1880-e0b0-11ea-899f-104e26c24507.png) ![image (1)](https://user-images.githubusercontent.com/12252921/90382000-c4de7280-e0b0-11ea-800a-a3d65a0185f6.png) ![image (2)](https://user-images.githubusercontent.com/12252921/90382002-c5770900-e0b0-11ea-9efd-c939c3f1a313.png) ![image (3)](https://user-images.githubusercontent.com/12252921/90382004-c6a83600-e0b0-11ea-8201-50deb55879ad.png) * Expected: When user switch data windows, the date in the top calendar should also sync with the data window * Actual: The date in the top calendar won't alter when switching data window by buttons **Logs** * <upload-error-log-here> **Environment** * mylog14-dashboard: v0.5.2-beta5
1.0
When user switch data window by "Today", "This Week" .. buttons, the calendar on top should also sync - **Steps to Reproduce** 1. User upload data to LogBoard 2. User switch the data window by the buttons "Today", "This Week"...... etc. **Results** ![image](https://user-images.githubusercontent.com/12252921/90381993-c27c1880-e0b0-11ea-899f-104e26c24507.png) ![image (1)](https://user-images.githubusercontent.com/12252921/90382000-c4de7280-e0b0-11ea-800a-a3d65a0185f6.png) ![image (2)](https://user-images.githubusercontent.com/12252921/90382002-c5770900-e0b0-11ea-9efd-c939c3f1a313.png) ![image (3)](https://user-images.githubusercontent.com/12252921/90382004-c6a83600-e0b0-11ea-8201-50deb55879ad.png) * Expected: When user switch data windows, the date in the top calendar should also sync with the data window * Actual: The date in the top calendar won't alter when switching data window by buttons **Logs** * <upload-error-log-here> **Environment** * mylog14-dashboard: v0.5.2-beta5
non_process
when user switch data window by today this week buttons the calendar on top should also sync steps to reproduce user upload data to logboard user switch the data window by the buttons today this week etc results expected when user switch data windows the date in the top calendar should also sync with the data window actual the date in the top calendar won t alter when switching data window by buttons logs environment dashboard
0
20,379
2,622,845,847
IssuesEvent
2015-03-04 08:02:31
max99x/pagemon-chrome-ext
https://api.github.com/repos/max99x/pagemon-chrome-ext
closed
Switch to the new favicon API
auto-migrated Priority-Low Type-Enhancement
``` Chrome 5+ exposes chrome://favicon/ URLs that can be used to get favicons by simply supplying a domain name. No longer do we need to explicitly save favicon URLs. ``` Original issue reported on code.google.com by `max99x` on 6 Jun 2010 at 3:31
1.0
Switch to the new favicon API - ``` Chrome 5+ exposes chrome://favicon/ URLs that can be used to get favicons by simply supplying a domain name. No longer do we need to explicitly save favicon URLs. ``` Original issue reported on code.google.com by `max99x` on 6 Jun 2010 at 3:31
non_process
switch to the new favicon api chrome exposes chrome favicon urls that can be used to get favicons by simply supplying a domain name no longer do we need to explicitly save favicon urls original issue reported on code google com by on jun at
0
20,958
27,817,251,759
IssuesEvent
2023-03-18 20:37:14
cse442-at-ub/project_s23-cinco
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
closed
Add routing to profile, account settings, and login button on navbar
Processing Task Sprint 2
*Task Tests* run in "Sprint2-Navbar-Buttons" branch in github test1: - in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser - look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"![navbar-profile-account-login.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/5e880ed6-8151-4e1e-927d-96291bebd670) -in a mobile view, click on the burger icon to ensure the following buttons are included: "profile, "login", "account settings":![mobile-view-buttons.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/42f79de8-067e-40d8-9c6e-3eb45821762a) -click on "profile" to get to edit profile page:![edit-profile-img.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/8e5f5415-76d1-4978-b4ec-302f1e0c321d) -click on "account settings" to get to the account settings page:![account-settings-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/a571095f-bd0c-4000-b3b8-5ed690986f0d) -click on "login" to get to login page:![login-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/ce877b30-5cea-4497-8a58-de3c627995b8)
1.0
Add routing to profile, account settings, and login button on navbar - *Task Tests* run in "Sprint2-Navbar-Buttons" branch in github test1: - in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser - look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"![navbar-profile-account-login.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/5e880ed6-8151-4e1e-927d-96291bebd670) -in a mobile view, click on the burger icon to ensure the following buttons are included: "profile, "login", "account settings":![mobile-view-buttons.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/42f79de8-067e-40d8-9c6e-3eb45821762a) -click on "profile" to get to edit profile page:![edit-profile-img.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/8e5f5415-76d1-4978-b4ec-302f1e0c321d) -click on "account settings" to get to the account settings page:![account-settings-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/a571095f-bd0c-4000-b3b8-5ed690986f0d) -click on "login" to get to login page:![login-page.PNG](https://images.zenhubusercontent.com/63e2d05608ee4c45e81f66f0/ce877b30-5cea-4497-8a58-de3c627995b8)
process
add routing to profile account settings and login button on navbar task tests run in navbar buttons branch in github in the project folder project cinco run npm start in the terminal this will open the homepage in a browser look at the navbar and ensure it has the following buttons profile account settings login in a mobile view click on the burger icon to ensure the following buttons are included profile login account settings click on profile to get to edit profile page click on account settings to get to the account settings page click on login to get to login page
1
4,623
4,489,147,864
IssuesEvent
2016-08-30 09:55:32
syl20bnr/spacemacs
https://api.github.com/repos/syl20bnr/spacemacs
closed
Spaceline performance bug?
- Bug tracker - Fixed upstream Performance
#### Description with Emacs 25.1.50.3 (built from source, git commit 8cfd9ba…, with -O2 and gtk toolkit on Linux Mint 17.3, using the spacemacs develop branch and freshly updated packages as of May 1st 2016), I experience extremely laggy cursor motion. Just writing a single line of text into an otherwise empty text-mode buffer and scrolling back and forth causes the cursor to stutter every ~20 characters. Disabling the modeline via t-m-t eliminates the issue; with spaceline enabled, I notice that I get on the order of 5 garbage collections for every line traversal (roughly 100 characters). Memory profiling shows that spaceline indeed allocates a surprisingly large amount of memory (this profile only shows the cost of 1 or 2 line traversals, `command-execute` down below was caused by helm when I toggled the profiler via `helm-M-x-read-extended-command`): ``` - redisplay_internal (C function) 19,812,999 70% - eval 19,812,999 70% - spaceline--prepare 19,350,659 68% - spaceline--prepare-any 18,402,191 65% - spaceline--eval-segment 14,209,815 50% + mapcar 6,775,993 24% + powerline-raw 1,064,653 3% + spaceline--eval-segment 320,340 1% + spaceline--segment-major-mode 127,352 0% + eval 5,024 0% + mapcar 1,389,412 4% + spaceline--get-face 565,504 2% + powerline-render 593,944 2% + powerline-width 349,444 1% powerline-fill 5,080 0% + command-execute 7,656,897 27% + winner-save-old-configurations 686,565 2% + evil-repeat-pre-hook 8,188 0% sp--save-pre-command-state 2,112 0% + command-error-default-function 2,032 0% internal-timer-start-idle 1,746 0% + timer-event-handler 1,572 0% ... 0 0% ``` For now, I'm working around the issue by setting the garbage collection threshold to 1GB and triggering collection after 5 idle seconds: ``` (setq gc-cons-threshold 1000000000) (run-with-idle-timer 5 t #'garbage-collect) ``` While this has made basic text navigation bearable again, I still find it a bit unsettling to know that spaceline is doing a lot of probably unnecessary work in the background. Could it be worthwhile to change spaceline from its current, nicely-functional style to a perhaps more efficient imperative style? Or is this performance problem completely unexpected and possibly caused by some configuration error on my side? #### Reproduction guide - Start Emacs - Move the cursor around #### System Info - OS: gnu/linux - Emacs: 25.1.50.3 - Spacemacs: 0.105.19 - Spacemacs branch: c-improved (rev. 1e67a11) (based on develop, with minor changes to c/c++-mode packages.el) - Graphic display: t - Distribution: spacemacs - Editing style: vim - Completion: helm - Layers: ```elisp (spacemacs-helm auto-completion (c-c++ :variables c-c++-enable-clang-support t c-c++-enable-irony-support t) colors context emacs-lisp ess evil-snipe (git :variables git-magit-status-fullscreen t) github gtags haskell html (latex :variables latex-enable-auto-fill nil) lua java javascript markdown org python ranger ruby (shell :variables shell-default-height 30 shell-default-position 'bottom) syntax-checking themes-megapack typescript typography (version-control :variables version-control-diff-tool 'git-gutter) (vinegar :variables vinegar-reuse-dired-buffer t) ycmd) ```
True
Spaceline performance bug? - #### Description with Emacs 25.1.50.3 (built from source, git commit 8cfd9ba…, with -O2 and gtk toolkit on Linux Mint 17.3, using the spacemacs develop branch and freshly updated packages as of May 1st 2016), I experience extremely laggy cursor motion. Just writing a single line of text into an otherwise empty text-mode buffer and scrolling back and forth causes the cursor to stutter every ~20 characters. Disabling the modeline via t-m-t eliminates the issue; with spaceline enabled, I notice that I get on the order of 5 garbage collections for every line traversal (roughly 100 characters). Memory profiling shows that spaceline indeed allocates a surprisingly large amount of memory (this profile only shows the cost of 1 or 2 line traversals, `command-execute` down below was caused by helm when I toggled the profiler via `helm-M-x-read-extended-command`): ``` - redisplay_internal (C function) 19,812,999 70% - eval 19,812,999 70% - spaceline--prepare 19,350,659 68% - spaceline--prepare-any 18,402,191 65% - spaceline--eval-segment 14,209,815 50% + mapcar 6,775,993 24% + powerline-raw 1,064,653 3% + spaceline--eval-segment 320,340 1% + spaceline--segment-major-mode 127,352 0% + eval 5,024 0% + mapcar 1,389,412 4% + spaceline--get-face 565,504 2% + powerline-render 593,944 2% + powerline-width 349,444 1% powerline-fill 5,080 0% + command-execute 7,656,897 27% + winner-save-old-configurations 686,565 2% + evil-repeat-pre-hook 8,188 0% sp--save-pre-command-state 2,112 0% + command-error-default-function 2,032 0% internal-timer-start-idle 1,746 0% + timer-event-handler 1,572 0% ... 0 0% ``` For now, I'm working around the issue by setting the garbage collection threshold to 1GB and triggering collection after 5 idle seconds: ``` (setq gc-cons-threshold 1000000000) (run-with-idle-timer 5 t #'garbage-collect) ``` While this has made basic text navigation bearable again, I still find it a bit unsettling to know that spaceline is doing a lot of probably unnecessary work in the background. Could it be worthwhile to change spaceline from its current, nicely-functional style to a perhaps more efficient imperative style? Or is this performance problem completely unexpected and possibly caused by some configuration error on my side? #### Reproduction guide - Start Emacs - Move the cursor around #### System Info - OS: gnu/linux - Emacs: 25.1.50.3 - Spacemacs: 0.105.19 - Spacemacs branch: c-improved (rev. 1e67a11) (based on develop, with minor changes to c/c++-mode packages.el) - Graphic display: t - Distribution: spacemacs - Editing style: vim - Completion: helm - Layers: ```elisp (spacemacs-helm auto-completion (c-c++ :variables c-c++-enable-clang-support t c-c++-enable-irony-support t) colors context emacs-lisp ess evil-snipe (git :variables git-magit-status-fullscreen t) github gtags haskell html (latex :variables latex-enable-auto-fill nil) lua java javascript markdown org python ranger ruby (shell :variables shell-default-height 30 shell-default-position 'bottom) syntax-checking themes-megapack typescript typography (version-control :variables version-control-diff-tool 'git-gutter) (vinegar :variables vinegar-reuse-dired-buffer t) ycmd) ```
non_process
spaceline performance bug description with emacs built from source git commit … with and gtk toolkit on linux mint using the spacemacs develop branch and freshly updated packages as of may i experience extremely laggy cursor motion just writing a single line of text into an otherwise empty text mode buffer and scrolling back and forth causes the cursor to stutter every characters disabling the modeline via t m t eliminates the issue with spaceline enabled i notice that i get on the order of garbage collections for every line traversal roughly characters memory profiling shows that spaceline indeed allocates a surprisingly large amount of memory this profile only shows the cost of or line traversals command execute down below was caused by helm when i toggled the profiler via helm m x read extended command redisplay internal c function eval spaceline prepare spaceline prepare any spaceline eval segment mapcar powerline raw spaceline eval segment spaceline segment major mode eval mapcar spaceline get face powerline render powerline width powerline fill command execute winner save old configurations evil repeat pre hook sp save pre command state command error default function internal timer start idle timer event handler for now i m working around the issue by setting the garbage collection threshold to and triggering collection after idle seconds setq gc cons threshold run with idle timer t garbage collect while this has made basic text navigation bearable again i still find it a bit unsettling to know that spaceline is doing a lot of probably unnecessary work in the background could it be worthwhile to change spaceline from its current nicely functional style to a perhaps more efficient imperative style or is this performance problem completely unexpected and possibly caused by some configuration error on my side reproduction guide start emacs move the cursor around system info os gnu linux emacs spacemacs spacemacs branch c improved rev based on develop with minor changes to c c mode packages el graphic display t distribution spacemacs editing style vim completion helm layers elisp spacemacs helm auto completion c c variables c c enable clang support t c c enable irony support t colors context emacs lisp ess evil snipe git variables git magit status fullscreen t github gtags haskell html latex variables latex enable auto fill nil lua java javascript markdown org python ranger ruby shell variables shell default height shell default position bottom syntax checking themes megapack typescript typography version control variables version control diff tool git gutter vinegar variables vinegar reuse dired buffer t ycmd
0
195,159
14,705,589,952
IssuesEvent
2021-01-04 18:23:34
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Number of secrets rewritten during encryption process is seen incorrect
[zube]: To Test kind/bug-qa
**What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - Deploy a cluster, and create some secrets - K get secrets -A --> On the cluster gives 137 secrets - enable secrets encryption ``` services: kube-api: secrets_encryption_config: enabled: true ``` - View in API of the cluster --> Do a rotateEncryptionKey action. - Logs of rancher server show that 150 secrets have been rewritten. - Logs in Rancher server: ``` 2020/12/17 04:28:54 [INFO] cluster [c-fzvn2] provisioning: Rewriting cluster secrets 2020/12/17 04:29:02 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 50 secrets rewritten 2020/12/17 04:29:14 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 100 secrets rewritten 2020/12/17 04:29:26 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 150 secrets rewritten 2020/12/17 04:29:29 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] Operation completed ``` **Expected Result:** The exact secrets rewritten must be shown. (maybe it is rounded off to the next batch size) **Other details that may be helpful:** Related to the fix for - https://github.com/rancher/rancher/issues/27735 **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.5-head - commit id: `bb69b03` - Installation option (single install/HA): Single node <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): RKE DO - Kubernetes version (use `kubectl version`): ``` 1.19 (Any) ```
1.0
Number of secrets rewritten during encryption process is seen incorrect - **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - Deploy a cluster, and create some secrets - K get secrets -A --> On the cluster gives 137 secrets - enable secrets encryption ``` services: kube-api: secrets_encryption_config: enabled: true ``` - View in API of the cluster --> Do a rotateEncryptionKey action. - Logs of rancher server show that 150 secrets have been rewritten. - Logs in Rancher server: ``` 2020/12/17 04:28:54 [INFO] cluster [c-fzvn2] provisioning: Rewriting cluster secrets 2020/12/17 04:29:02 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 50 secrets rewritten 2020/12/17 04:29:14 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 100 secrets rewritten 2020/12/17 04:29:26 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 150 secrets rewritten 2020/12/17 04:29:29 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] Operation completed ``` **Expected Result:** The exact secrets rewritten must be shown. (maybe it is rounded off to the next batch size) **Other details that may be helpful:** Related to the fix for - https://github.com/rancher/rancher/issues/27735 **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.5-head - commit id: `bb69b03` - Installation option (single install/HA): Single node <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): RKE DO - Kubernetes version (use `kubectl version`): ``` 1.19 (Any) ```
non_process
number of secrets rewritten during encryption process is seen incorrect what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible deploy a cluster and create some secrets k get secrets a on the cluster gives secrets enable secrets encryption services kube api secrets encryption config enabled true view in api of the cluster do a rotateencryptionkey action logs of rancher server show that secrets have been rewritten logs in rancher server cluster provisioning rewriting cluster secrets cluster provisioning secrets rewritten cluster provisioning secrets rewritten cluster provisioning secrets rewritten cluster provisioning operation completed expected result the exact secrets rewritten must be shown maybe it is rounded off to the next batch size other details that may be helpful related to the fix for environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui head commit id installation option single install ha single node if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported rke do kubernetes version use kubectl version any
0
284,719
21,466,688,488
IssuesEvent
2022-04-26 05:02:22
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Docs] #13254 [Bug]-[800]:Close button can be clicked on a disabled select widget
Documentation User Education Pod
> TODO - [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket - [ ] Fill these fields - [ ] Prepare first draft - [ ] Add label: "Ready for Docs Team" Field | Details -----|----- **POD** | App Viewers Pod **Parent Ticket** | #13254 Engineer | Release Date | Live Date | First Draft | Auto Assign | Priority | Environment |
1.0
[Docs] #13254 [Bug]-[800]:Close button can be clicked on a disabled select widget - > TODO - [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket - [ ] Fill these fields - [ ] Prepare first draft - [ ] Add label: "Ready for Docs Team" Field | Details -----|----- **POD** | App Viewers Pod **Parent Ticket** | #13254 Engineer | Release Date | Live Date | First Draft | Auto Assign | Priority | Environment |
non_process
close button can be clicked on a disabled select widget todo evaluate if this task is needed if not add the skip docs label on the parent ticket fill these fields prepare first draft add label ready for docs team field details pod app viewers pod parent ticket engineer release date live date first draft auto assign priority environment
0
136,102
30,475,790,514
IssuesEvent
2023-07-17 16:24:23
ita-social-projects/StreetCode
https://api.github.com/repos/ita-social-projects/StreetCode
opened
[Admin/History map] The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block
bug (Epic#2) Admin/New StreetCode
**Environment:** OS: Windows 10 Pro Browser: Google Chrome Version 111.0.5563.112. **Reproducible:** always. **Build found:** commit [d494c37](https://github.com/ita-social-projects/StreetCode/commit/d494c372c230bf30fef322fdb50405a1c708c55b) **Type:** Functional **Priority:** Low **Severity:** Low **Preconditions** 1. Go to the site. 2. Login as admin. 3. Open the new StreetCode page or the StreetCode page for editing. **Steps to reproduce:** 1. Go to the "Мапа історії" block. 2. Enter the address of location the streetcode to the "Додати стріткод на мапу" form or click on the desired location on the map. 3. Enter the number of the streetcode plate with a negative value. 4. Pay atention to the information about the location of the streetcode under the map. **Actual result** The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block. **Expected result** The system restricts adding the number of the streetcode plate with a negative value. **User story and test case links** User story #126 Test case #362 <img width="1125" alt="number" src="https://github.com/ita-social-projects/StreetCode/assets/135837034/6578f8ec-35fe-4837-a740-5aabb318b228">
1.0
[Admin/History map] The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block - **Environment:** OS: Windows 10 Pro Browser: Google Chrome Version 111.0.5563.112. **Reproducible:** always. **Build found:** commit [d494c37](https://github.com/ita-social-projects/StreetCode/commit/d494c372c230bf30fef322fdb50405a1c708c55b) **Type:** Functional **Priority:** Low **Severity:** Low **Preconditions** 1. Go to the site. 2. Login as admin. 3. Open the new StreetCode page or the StreetCode page for editing. **Steps to reproduce:** 1. Go to the "Мапа історії" block. 2. Enter the address of location the streetcode to the "Додати стріткод на мапу" form or click on the desired location on the map. 3. Enter the number of the streetcode plate with a negative value. 4. Pay atention to the information about the location of the streetcode under the map. **Actual result** The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block. **Expected result** The system restricts adding the number of the streetcode plate with a negative value. **User story and test case links** User story #126 Test case #362 <img width="1125" alt="number" src="https://github.com/ita-social-projects/StreetCode/assets/135837034/6578f8ec-35fe-4837-a740-5aabb318b228">
non_process
the number of the streetcode plate with a negative value is added to the map in the мапа історії block environment os windows pro browser google chrome version reproducible always build found commit type functional priority low severity low preconditions go to the site login as admin open the new streetcode page or the streetcode page for editing steps to reproduce go to the мапа історії block enter the address of location the streetcode to the додати стріткод на мапу form or click on the desired location on the map enter the number of the streetcode plate with a negative value pay atention to the information about the location of the streetcode under the map actual result the number of the streetcode plate with a negative value is added to the map in the мапа історії block expected result the system restricts adding the number of the streetcode plate with a negative value user story and test case links user story test case img width alt number src
0
16,784
21,970,817,600
IssuesEvent
2022-05-25 03:25:18
quark-engine/quark-engine
https://api.github.com/repos/quark-engine/quark-engine
closed
Update author information
issue-processing-state-06
Since the main author was changed, we should update the author information in setup.py.
1.0
Update author information - Since the main author was changed, we should update the author information in setup.py.
process
update author information since the main author was changed we should update the author information in setup py
1
3,589
6,621,672,102
IssuesEvent
2017-09-21 20:06:36
WikiWatershed/rapid-watershed-delineation
https://api.github.com/repos/WikiWatershed/rapid-watershed-delineation
closed
Expose unsimplified result
BigCZ Geoprocessing API
Provide a way for users to get the unsimplified geojson results. The MMW Geoprocessing API would like to expose them.
1.0
Expose unsimplified result - Provide a way for users to get the unsimplified geojson results. The MMW Geoprocessing API would like to expose them.
process
expose unsimplified result provide a way for users to get the unsimplified geojson results the mmw geoprocessing api would like to expose them
1
19,591
25,932,202,326
IssuesEvent
2022-12-16 10:59:19
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Node hangs when running a stand-alone script with captured output
help wanted windows process
* **Version**: v8.1.3 * **Platform**: Windows 10 version 10.0.14393 x64 * **Subsystem**: I am running into an issue on our build servers where Node hangs after running a stand-alone script if it is launched with its standard handles redirected to pipes. Specifically, I am using Ruby's `popen3` function, which wraps OS pipe functionality to obtain pipe streams and then passes the underlying handles to the C runtime library's `popen` API function. I don't _believe_ there is anything particular about the way Ruby is doing this, and I have the expectation that similar code written in C++ directly calling the underlying API functions would exhibit the same problem. I have not actually tested this, though. I first observed this behaviour as part of an `npm install`. When NPM installs the `node-sass` module, it tries to run an install hook script, and the resulting `node scripts/install.js` process hangs until it is sent a break signal or terminated externally. This does not occur if `npm install` is simply run at the command-line, nor does it occur with simple piping done via the command interpreter (`npm install | more`, for instance). In further debugging, I tried installing a local copy of `node-sass` in which I made alterations to the `scripts/install.js` script. The hang is reproducible even if the entire `install.js` script is a single `console.log` line. Furthermore, if I outright remove the `scripts/install.js` file, then the parent process of what would have been `node scripts/install.js` encounters an error (expected), complains loudly about it, and then hangs. The command-line of this parent process is: C:\Program Files\nodejs\\node.exe "C:\NPM\node_modules\npm\bin\npm-cli.js" install "C:\Path\To\node-sass" I'm not sure what the next step is in identifying the root cause of this issue. I suspect NPM might play some part in it, because if I try my test on a dummy script file _directly_ (`node test.js`), then the hang is not observed. The following Ruby code should demonstrate the hang if the `node-sass` module is configured to be installed into `node_modules` in the current project directory: require 'open3' Open3.popen3("npm install node-sass") do |stdin, stdout, stderr, thr| stdout_thread = Thread.new { stdout.each_line { |line| puts line } } stderr_thread = Thread.new { stderr.each_line { |line| puts "[STDERR] #{line}" } } stdout_thread.join stderr_thread.join puts "Exit code: #{thr.value.exitstatus}" end At the time of the hang, if I attach a debugger to the rogue `node.exe` process, I see the following threads: Main thread: => Blocked in OS API: GetQueuedCompletionStatusEx => Call stack: node.exe!uv_poll_ex(uv_loop_s * loop, unsigned long timeout) Line 432 C node.exe!uv_run(uv_loop_s * loop, uv_run_mode mode) Line 513 C node.exe!node::Start(v8::Isolate * isolate, node::IsolateData * isolate_data, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4520 C++ node.exe!node::Start(uv_loop_s * event_loop, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4585 C++ node.exe!node::Start(int argc, char * * argv) Line 4640 C++ [Inline Frame] node.exe!invoke_main() Line 79 C++ Worker threads: 8 threads are in this state => Blocked in OS API: WaitForSingleObjectEx => Call stack: [Inline Frame] node.exe!v8::base::Semaphore::Wait() Line 159 C++ [Inline Frame] node.exe!v8::platform::TaskQueue::GetNext() Line 46 C++ node.exe!v8::platform::WorkerThread::Run() Line 25 C++ [Inline Frame] node.exe!v8::base::Thread::NotifyStartedAndRun() Line 512 C++ node.exe!v8::base::ThreadEntry(void * arg) Line 1328 C++ [Inline Frame] node.exe!invoke_thread_procedure(unsigned int(*)(void *) context, void * const) Line 91 C++ node.exe!thread_start<unsigned int (__cdecl*)(void * __ptr64)>(void * const parameter) Line 115 C++ Thread pool threads: 4 threads are in this state => Blocked in OS API: NtWaitForWorkViaWorkerFactory => Call stack: ntdll.dll!TppWorkerThread() Note that the process that is hanging is _not_ the direct descendant of the calling site that is waiting (in this case the Ruby interpreter). The child process launched by `popen3` has in fact exited, but the _streams_ (`stdour`, `stderr`) aren't signalled because they were inherited by grandchildren, and one of those is the culprit that isn't exiting. [WAITING] Ruby code: `popen "npm install"` `- [exited] cmd.exe npm.bat |- [exited] node.exe npm-cli.js prefix -g `- [exited] node.exe npm-cli.js install `- [exited] cmd.exe /d /s /c node.exe scripts/install.js `- [STILL RUNNING/HUNG] node.exe scripts/install.js The hung process is sitting around keeping `stdout` and `stderr` open, and since the pattern of the call site is to wait for the end of those streams as the indicator that the child process tree is complete, the call site blocks indefinitely as well. I have experimented with altering the call site so that it instead explicitly waits for its direct child process to exit, and forces the streams closed. This does permit the calling code to proceed, but the rogue process still remains, which (since this is Windows) causes issues later on to do with file locking. If there is further debugging I can do to try to narrow down the cause of this issue, I would appreciate any advice. :-) Thanks!
1.0
Node hangs when running a stand-alone script with captured output - * **Version**: v8.1.3 * **Platform**: Windows 10 version 10.0.14393 x64 * **Subsystem**: I am running into an issue on our build servers where Node hangs after running a stand-alone script if it is launched with its standard handles redirected to pipes. Specifically, I am using Ruby's `popen3` function, which wraps OS pipe functionality to obtain pipe streams and then passes the underlying handles to the C runtime library's `popen` API function. I don't _believe_ there is anything particular about the way Ruby is doing this, and I have the expectation that similar code written in C++ directly calling the underlying API functions would exhibit the same problem. I have not actually tested this, though. I first observed this behaviour as part of an `npm install`. When NPM installs the `node-sass` module, it tries to run an install hook script, and the resulting `node scripts/install.js` process hangs until it is sent a break signal or terminated externally. This does not occur if `npm install` is simply run at the command-line, nor does it occur with simple piping done via the command interpreter (`npm install | more`, for instance). In further debugging, I tried installing a local copy of `node-sass` in which I made alterations to the `scripts/install.js` script. The hang is reproducible even if the entire `install.js` script is a single `console.log` line. Furthermore, if I outright remove the `scripts/install.js` file, then the parent process of what would have been `node scripts/install.js` encounters an error (expected), complains loudly about it, and then hangs. The command-line of this parent process is: C:\Program Files\nodejs\\node.exe "C:\NPM\node_modules\npm\bin\npm-cli.js" install "C:\Path\To\node-sass" I'm not sure what the next step is in identifying the root cause of this issue. I suspect NPM might play some part in it, because if I try my test on a dummy script file _directly_ (`node test.js`), then the hang is not observed. The following Ruby code should demonstrate the hang if the `node-sass` module is configured to be installed into `node_modules` in the current project directory: require 'open3' Open3.popen3("npm install node-sass") do |stdin, stdout, stderr, thr| stdout_thread = Thread.new { stdout.each_line { |line| puts line } } stderr_thread = Thread.new { stderr.each_line { |line| puts "[STDERR] #{line}" } } stdout_thread.join stderr_thread.join puts "Exit code: #{thr.value.exitstatus}" end At the time of the hang, if I attach a debugger to the rogue `node.exe` process, I see the following threads: Main thread: => Blocked in OS API: GetQueuedCompletionStatusEx => Call stack: node.exe!uv_poll_ex(uv_loop_s * loop, unsigned long timeout) Line 432 C node.exe!uv_run(uv_loop_s * loop, uv_run_mode mode) Line 513 C node.exe!node::Start(v8::Isolate * isolate, node::IsolateData * isolate_data, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4520 C++ node.exe!node::Start(uv_loop_s * event_loop, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4585 C++ node.exe!node::Start(int argc, char * * argv) Line 4640 C++ [Inline Frame] node.exe!invoke_main() Line 79 C++ Worker threads: 8 threads are in this state => Blocked in OS API: WaitForSingleObjectEx => Call stack: [Inline Frame] node.exe!v8::base::Semaphore::Wait() Line 159 C++ [Inline Frame] node.exe!v8::platform::TaskQueue::GetNext() Line 46 C++ node.exe!v8::platform::WorkerThread::Run() Line 25 C++ [Inline Frame] node.exe!v8::base::Thread::NotifyStartedAndRun() Line 512 C++ node.exe!v8::base::ThreadEntry(void * arg) Line 1328 C++ [Inline Frame] node.exe!invoke_thread_procedure(unsigned int(*)(void *) context, void * const) Line 91 C++ node.exe!thread_start<unsigned int (__cdecl*)(void * __ptr64)>(void * const parameter) Line 115 C++ Thread pool threads: 4 threads are in this state => Blocked in OS API: NtWaitForWorkViaWorkerFactory => Call stack: ntdll.dll!TppWorkerThread() Note that the process that is hanging is _not_ the direct descendant of the calling site that is waiting (in this case the Ruby interpreter). The child process launched by `popen3` has in fact exited, but the _streams_ (`stdour`, `stderr`) aren't signalled because they were inherited by grandchildren, and one of those is the culprit that isn't exiting. [WAITING] Ruby code: `popen "npm install"` `- [exited] cmd.exe npm.bat |- [exited] node.exe npm-cli.js prefix -g `- [exited] node.exe npm-cli.js install `- [exited] cmd.exe /d /s /c node.exe scripts/install.js `- [STILL RUNNING/HUNG] node.exe scripts/install.js The hung process is sitting around keeping `stdout` and `stderr` open, and since the pattern of the call site is to wait for the end of those streams as the indicator that the child process tree is complete, the call site blocks indefinitely as well. I have experimented with altering the call site so that it instead explicitly waits for its direct child process to exit, and forces the streams closed. This does permit the calling code to proceed, but the rogue process still remains, which (since this is Windows) causes issues later on to do with file locking. If there is further debugging I can do to try to narrow down the cause of this issue, I would appreciate any advice. :-) Thanks!
process
node hangs when running a stand alone script with captured output version platform windows version subsystem i am running into an issue on our build servers where node hangs after running a stand alone script if it is launched with its standard handles redirected to pipes specifically i am using ruby s function which wraps os pipe functionality to obtain pipe streams and then passes the underlying handles to the c runtime library s popen api function i don t believe there is anything particular about the way ruby is doing this and i have the expectation that similar code written in c directly calling the underlying api functions would exhibit the same problem i have not actually tested this though i first observed this behaviour as part of an npm install when npm installs the node sass module it tries to run an install hook script and the resulting node scripts install js process hangs until it is sent a break signal or terminated externally this does not occur if npm install is simply run at the command line nor does it occur with simple piping done via the command interpreter npm install more for instance in further debugging i tried installing a local copy of node sass in which i made alterations to the scripts install js script the hang is reproducible even if the entire install js script is a single console log line furthermore if i outright remove the scripts install js file then the parent process of what would have been node scripts install js encounters an error expected complains loudly about it and then hangs the command line of this parent process is c program files nodejs node exe c npm node modules npm bin npm cli js install c path to node sass i m not sure what the next step is in identifying the root cause of this issue i suspect npm might play some part in it because if i try my test on a dummy script file directly node test js then the hang is not observed the following ruby code should demonstrate the hang if the node sass module is configured to be installed into node modules in the current project directory require npm install node sass do stdin stdout stderr thr stdout thread thread new stdout each line line puts line stderr thread thread new stderr each line line puts line stdout thread join stderr thread join puts exit code thr value exitstatus end at the time of the hang if i attach a debugger to the rogue node exe process i see the following threads main thread blocked in os api getqueuedcompletionstatusex call stack node exe uv poll ex uv loop s loop unsigned long timeout line c node exe uv run uv loop s loop uv run mode mode line c node exe node start isolate isolate node isolatedata isolate data int argc const char const argv int exec argc const char const exec argv line c node exe node start uv loop s event loop int argc const char const argv int exec argc const char const exec argv line c node exe node start int argc char argv line c node exe invoke main line c worker threads threads are in this state blocked in os api waitforsingleobjectex call stack node exe base semaphore wait line c node exe platform taskqueue getnext line c node exe platform workerthread run line c node exe base thread notifystartedandrun line c node exe base threadentry void arg line c node exe invoke thread procedure unsigned int void context void const line c node exe thread start void const parameter line c thread pool threads threads are in this state blocked in os api ntwaitforworkviaworkerfactory call stack ntdll dll tppworkerthread note that the process that is hanging is not the direct descendant of the calling site that is waiting in this case the ruby interpreter the child process launched by has in fact exited but the streams stdour stderr aren t signalled because they were inherited by grandchildren and one of those is the culprit that isn t exiting ruby code popen npm install cmd exe npm bat node exe npm cli js prefix g node exe npm cli js install cmd exe d s c node exe scripts install js node exe scripts install js the hung process is sitting around keeping stdout and stderr open and since the pattern of the call site is to wait for the end of those streams as the indicator that the child process tree is complete the call site blocks indefinitely as well i have experimented with altering the call site so that it instead explicitly waits for its direct child process to exit and forces the streams closed this does permit the calling code to proceed but the rogue process still remains which since this is windows causes issues later on to do with file locking if there is further debugging i can do to try to narrow down the cause of this issue i would appreciate any advice thanks
1
288,708
8,850,587,584
IssuesEvent
2019-01-08 13:41:30
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
jobs.jobvite.com - design is broken
browser-firefox priority-normal
<!-- @browser: Firefox 65.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:65.0) Gecko/20100101 Firefox/65.0 --> <!-- @reported_with: desktop-reporter --> **URL**: http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed **Browser / Version**: Firefox 65.0 **Operating System**: Linux **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Site will not submit application **Steps to Reproduce**: Last button does not function. [![Screenshot Description](https://webcompat.com/uploads/2018/12/eddec568-ea29-42c3-bcad-aca71a843db1-thumb.jpeg)](https://webcompat.com/uploads/2018/12/eddec568-ea29-42c3-bcad-aca71a843db1.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181207174328</li><li>tracking content blocked: true (strict)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "unreachable code after return statement" {file: "http://app.jvicdn.com/__assets__/lib/moment/moment.min.js?v=181213135315" line: 6 column: 12432 source: "e._d.setTime(+e._d+f),a.updateOffset(e,!1),e):Aa(b).local();return c._isUTC?Aa(b).zone(c._offset||0):Aa(b).local()}funct"}]', u'[JavaScript Warning: "The resource at http://www.google-analytics.com/ga.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://www.google-analytics.com/ga.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 35}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.linkedin.com/in.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 346}]', u'[JavaScript Error: "The stylesheet http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed was not loaded because its MIME type, text/html, is not text/css." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "The resource at http://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.twitter.com/widgets.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[JavaScript Warning: "The resource at http://connect.facebook.net/en_US/sdk.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://connect.facebook.net/en_US/sdk.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[console.log(Replacing social widget for LinkedIn) moz-extension://921d6c32-4cd8-46d4-93b8-c67e37371904/js/contentscripts/socialwidgets.js:306:5]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "moz-extension://381126cf-dbe6-450d-8097-a68a2e0a1032/content_scripts/handler.js" line: 73}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
jobs.jobvite.com - design is broken - <!-- @browser: Firefox 65.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:65.0) Gecko/20100101 Firefox/65.0 --> <!-- @reported_with: desktop-reporter --> **URL**: http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed **Browser / Version**: Firefox 65.0 **Operating System**: Linux **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Site will not submit application **Steps to Reproduce**: Last button does not function. [![Screenshot Description](https://webcompat.com/uploads/2018/12/eddec568-ea29-42c3-bcad-aca71a843db1-thumb.jpeg)](https://webcompat.com/uploads/2018/12/eddec568-ea29-42c3-bcad-aca71a843db1.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181207174328</li><li>tracking content blocked: true (strict)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "unreachable code after return statement" {file: "http://app.jvicdn.com/__assets__/lib/moment/moment.min.js?v=181213135315" line: 6 column: 12432 source: "e._d.setTime(+e._d+f),a.updateOffset(e,!1),e):Aa(b).local();return c._isUTC?Aa(b).zone(c._offset||0):Aa(b).local()}funct"}]', u'[JavaScript Warning: "The resource at http://www.google-analytics.com/ga.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://www.google-analytics.com/ga.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 35}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.linkedin.com/in.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 346}]', u'[JavaScript Error: "The stylesheet http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed was not loaded because its MIME type, text/html, is not text/css." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "The resource at http://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.twitter.com/widgets.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[JavaScript Warning: "The resource at http://connect.facebook.net/en_US/sdk.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://connect.facebook.net/en_US/sdk.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[console.log(Replacing social widget for LinkedIn) moz-extension://921d6c32-4cd8-46d4-93b8-c67e37371904/js/contentscripts/socialwidgets.js:306:5]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "moz-extension://381126cf-dbe6-450d-8097-a68a2e0a1032/content_scripts/handler.js" line: 73}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
jobs jobvite com design is broken url browser version firefox operating system linux tested another browser yes problem type design is broken description site will not submit application steps to reproduce last button does not function browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked true strict gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel aurora console messages u u u u u u u u u u from with ❤️
0
674,133
23,040,439,666
IssuesEvent
2022-07-23 04:00:40
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
Shuttle/vehicle movement structure seems dodgy
Priority: 3-Not Required Needs Discussion Difficulty: 3-Hard
## Description This issue is very much an abstract code-structure issue right now, but I think it'll be a problem later. I was told to open an issue on this. Right now, there is no actual concept of movement states (i.e. no interface for such). However, there are three movement modes regardless. They are laid out in an order which is not respective of priority in code that otherwise seems like it's supposed to use this information to achieve prioritization (server `MoverController`): + Mob Movement - Always applied if `IMobMoverComponent` is available. Adds to `_excludedMobs`, does not read. Standard movement. Cares about CanMove (but this is 'interior', not relevant for mode prioritization) + Shuttle Movement - Applied if `PilotComponent` is available *and* it is connected to a console. Adds to `_excludedMobs`, does not read. Shuttle control. + Kinematic Movement - Always applied if `IMoverComponent` (required for any movement processing whatsoever) is available. Reads and ignores those entities in `_excludedMobs`. Ghost movement - ignores CanMove, etc. This points to some attempt to use `_excludedMobs` to use prioritization to keep the states consistent, but this wasn't actually made use of properly, so instead `PilotComponent` is also a marker that enacts a movement block if it's connected to a console. ## Specific concerns The specific concerns I have with this approach are that adding states to this will inevitably require those states trigger movement blocks, which will prevent *other* movement blocks from being tested for even when they really should be. (In particular, I was looking into how this might have an effect on the development of ~~clown cars~~ non-shuttle vehicles. The answer appears to be that to implement those, one would need to either handle it via buckling (likely not an option for various reasons under normal circumstances) or implement some sort of mirror system to piloting for the driver of a vehicle, doing a movement block, etc. The worry I have is that doing such a thing would interact badly with if someone driving a vehicle then went on to attempt to pilot a shuttle while driving a vehicle. In this case, both movement states would then be in effect simultaneously.) ## Citations Relevant code (as of master at time of writing) with their roles in shuttle control is here: + Ignores shuttle control but gets disabled by movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Client/Physics/Controllers/MoverController.cs + Contains the action blocker call which causes movement attempt events: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Movement/SharedMoverController.cs + The actual server-side authoritative movement logic, including the priority-disable of kinematic movement: https://github.com/space-wizards/space-station-14/blob/master/Content.Server/Physics/Controllers/MoverController.cs + Shuttle movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Shuttles/SharedShuttleConsoleSystem.cs
1.0
Shuttle/vehicle movement structure seems dodgy - ## Description This issue is very much an abstract code-structure issue right now, but I think it'll be a problem later. I was told to open an issue on this. Right now, there is no actual concept of movement states (i.e. no interface for such). However, there are three movement modes regardless. They are laid out in an order which is not respective of priority in code that otherwise seems like it's supposed to use this information to achieve prioritization (server `MoverController`): + Mob Movement - Always applied if `IMobMoverComponent` is available. Adds to `_excludedMobs`, does not read. Standard movement. Cares about CanMove (but this is 'interior', not relevant for mode prioritization) + Shuttle Movement - Applied if `PilotComponent` is available *and* it is connected to a console. Adds to `_excludedMobs`, does not read. Shuttle control. + Kinematic Movement - Always applied if `IMoverComponent` (required for any movement processing whatsoever) is available. Reads and ignores those entities in `_excludedMobs`. Ghost movement - ignores CanMove, etc. This points to some attempt to use `_excludedMobs` to use prioritization to keep the states consistent, but this wasn't actually made use of properly, so instead `PilotComponent` is also a marker that enacts a movement block if it's connected to a console. ## Specific concerns The specific concerns I have with this approach are that adding states to this will inevitably require those states trigger movement blocks, which will prevent *other* movement blocks from being tested for even when they really should be. (In particular, I was looking into how this might have an effect on the development of ~~clown cars~~ non-shuttle vehicles. The answer appears to be that to implement those, one would need to either handle it via buckling (likely not an option for various reasons under normal circumstances) or implement some sort of mirror system to piloting for the driver of a vehicle, doing a movement block, etc. The worry I have is that doing such a thing would interact badly with if someone driving a vehicle then went on to attempt to pilot a shuttle while driving a vehicle. In this case, both movement states would then be in effect simultaneously.) ## Citations Relevant code (as of master at time of writing) with their roles in shuttle control is here: + Ignores shuttle control but gets disabled by movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Client/Physics/Controllers/MoverController.cs + Contains the action blocker call which causes movement attempt events: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Movement/SharedMoverController.cs + The actual server-side authoritative movement logic, including the priority-disable of kinematic movement: https://github.com/space-wizards/space-station-14/blob/master/Content.Server/Physics/Controllers/MoverController.cs + Shuttle movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Shuttles/SharedShuttleConsoleSystem.cs
non_process
shuttle vehicle movement structure seems dodgy description this issue is very much an abstract code structure issue right now but i think it ll be a problem later i was told to open an issue on this right now there is no actual concept of movement states i e no interface for such however there are three movement modes regardless they are laid out in an order which is not respective of priority in code that otherwise seems like it s supposed to use this information to achieve prioritization server movercontroller mob movement always applied if imobmovercomponent is available adds to excludedmobs does not read standard movement cares about canmove but this is interior not relevant for mode prioritization shuttle movement applied if pilotcomponent is available and it is connected to a console adds to excludedmobs does not read shuttle control kinematic movement always applied if imovercomponent required for any movement processing whatsoever is available reads and ignores those entities in excludedmobs ghost movement ignores canmove etc this points to some attempt to use excludedmobs to use prioritization to keep the states consistent but this wasn t actually made use of properly so instead pilotcomponent is also a marker that enacts a movement block if it s connected to a console specific concerns the specific concerns i have with this approach are that adding states to this will inevitably require those states trigger movement blocks which will prevent other movement blocks from being tested for even when they really should be in particular i was looking into how this might have an effect on the development of clown cars non shuttle vehicles the answer appears to be that to implement those one would need to either handle it via buckling likely not an option for various reasons under normal circumstances or implement some sort of mirror system to piloting for the driver of a vehicle doing a movement block etc the worry i have is that doing such a thing would interact badly with if someone driving a vehicle then went on to attempt to pilot a shuttle while driving a vehicle in this case both movement states would then be in effect simultaneously citations relevant code as of master at time of writing with their roles in shuttle control is here ignores shuttle control but gets disabled by movement blocker contains the action blocker call which causes movement attempt events the actual server side authoritative movement logic including the priority disable of kinematic movement shuttle movement blocker
0
13,506
16,045,207,919
IssuesEvent
2021-04-22 12:55:48
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Prisma Format: Fill in native type attributes on foreign key relations
kind/bug process/candidate team/client topic: mongodb
## Problem This problem popped up while working with Mongo. Given the following: ```prisma datasource db { provider = "mongodb" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" previewFeatures = ["mongodb"] } model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String } ``` All good so far, when we go to link these two models with User posts: ```prisma model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String posts Post[] } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String } ``` It auto-formats to: ```prisma model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String posts Post[] } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String User User? @relation(fields: [userId], references: [id]) userId String? } ``` There's a red line under `User User?` with the following: ``` Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`. ``` You can also see this when you run `npx prisma format`: ``` Environment variables loaded from .env Prisma schema loaded from prisma/schema.prisma Error: Schema parsing error: Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`. --> schema.prisma:21 | 21 | User User? @relation(fields: [userId], references: [id]) 22 | userId String? | Validation Error Count: 1 ``` This is slightly opaque, but the solution is to add `@db.ObjectId` to `userId` like so: ```prisma model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String posts Post[] } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String User User? @relation(fields: [userId], references: [id]) userId String? @db.ObjectId } ```
1.0
Prisma Format: Fill in native type attributes on foreign key relations - ## Problem This problem popped up while working with Mongo. Given the following: ```prisma datasource db { provider = "mongodb" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" previewFeatures = ["mongodb"] } model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String } ``` All good so far, when we go to link these two models with User posts: ```prisma model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String posts Post[] } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String } ``` It auto-formats to: ```prisma model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String posts Post[] } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String User User? @relation(fields: [userId], references: [id]) userId String? } ``` There's a red line under `User User?` with the following: ``` Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`. ``` You can also see this when you run `npx prisma format`: ``` Environment variables loaded from .env Prisma schema loaded from prisma/schema.prisma Error: Schema parsing error: Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`. --> schema.prisma:21 | 21 | User User? @relation(fields: [userId], references: [id]) 22 | userId String? | Validation Error Count: 1 ``` This is slightly opaque, but the solution is to add `@db.ObjectId` to `userId` like so: ```prisma model User { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId name String posts Post[] } model Post { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId title String User User? @relation(fields: [userId], references: [id]) userId String? @db.ObjectId } ```
process
prisma format fill in native type attributes on foreign key relations problem this problem popped up while working with mongo given the following prisma datasource db provider mongodb url env database url generator client provider prisma client js previewfeatures model user id string id default dbgenerated map id db objectid name string model post id string id default dbgenerated map id db objectid title string all good so far when we go to link these two models with user posts prisma model user id string id default dbgenerated map id db objectid name string posts post model post id string id default dbgenerated map id db objectid title string it auto formats to prisma model user id string id default dbgenerated map id db objectid name string posts post model post id string id default dbgenerated map id db objectid title string user user relation fields references userid string there s a red line under user user with the following error parsing attribute relation the type of the field userid in the model post is not matching the type of the referenced field id in model user you can also see this when you run npx prisma format environment variables loaded from env prisma schema loaded from prisma schema prisma error schema parsing error error parsing attribute relation the type of the field userid in the model post is not matching the type of the referenced field id in model user schema prisma user user relation fields references userid string validation error count this is slightly opaque but the solution is to add db objectid to userid like so prisma model user id string id default dbgenerated map id db objectid name string posts post model post id string id default dbgenerated map id db objectid title string user user relation fields references userid string db objectid
1
212,504
16,487,440,007
IssuesEvent
2021-05-24 20:16:25
anitab-org/stem-diverse-tv
https://api.github.com/repos/anitab-org/stem-diverse-tv
closed
Bug : Incorrect link to bug_report.md in contributing_guidelines.md
Category: Documentation/Training First Timers Only Type: Bug
### Describe the bug The links to bug report is incorrect in the Contributing Guidelines file you can see this by also scrolling to the [Contributing section](https://github.com/anitab-org/stem-diverse-tv#contributing) in readme ### To Reproduce Steps to reproduce the behavior: 1. Go to 'https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/contributing_guidelines.md' 2. Scroll down to Contribution guidelines section 3. Click on [bug_report.md](https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/ISSUE_TEMPLATE/bug_report.md) 4. See error ### Expected behavior Links to the file should be corrected bug report: https://github.com/anitab-org/stem-diverse-tv/blob/master/.github/ISSUE_TEMPLATE/bug_report.md ### Desktop (please complete the following information): - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ### Smartphone (please complete the following information): - Device: [e.g. iPhone6] - OS: [e.g. iOS8.1] - Browser [e.g. stock browser, safari] - Version [e.g. 22] ### Additional context Add any other context about the problem here.
1.0
Bug : Incorrect link to bug_report.md in contributing_guidelines.md - ### Describe the bug The links to bug report is incorrect in the Contributing Guidelines file you can see this by also scrolling to the [Contributing section](https://github.com/anitab-org/stem-diverse-tv#contributing) in readme ### To Reproduce Steps to reproduce the behavior: 1. Go to 'https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/contributing_guidelines.md' 2. Scroll down to Contribution guidelines section 3. Click on [bug_report.md](https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/ISSUE_TEMPLATE/bug_report.md) 4. See error ### Expected behavior Links to the file should be corrected bug report: https://github.com/anitab-org/stem-diverse-tv/blob/master/.github/ISSUE_TEMPLATE/bug_report.md ### Desktop (please complete the following information): - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ### Smartphone (please complete the following information): - Device: [e.g. iPhone6] - OS: [e.g. iOS8.1] - Browser [e.g. stock browser, safari] - Version [e.g. 22] ### Additional context Add any other context about the problem here.
non_process
bug incorrect link to bug report md in contributing guidelines md describe the bug the links to bug report is incorrect in the contributing guidelines file you can see this by also scrolling to the in readme to reproduce steps to reproduce the behavior go to scroll down to contribution guidelines section click on see error expected behavior links to the file should be corrected bug report desktop please complete the following information os browser version smartphone please complete the following information device os browser version additional context add any other context about the problem here
0
344,704
24,824,023,485
IssuesEvent
2022-10-25 18:56:29
pnp/powershell
https://api.github.com/repos/pnp/powershell
closed
Documentation: Using pnp/powershell Docker Image
documentation
**Is your feature request related to a problem? Please describe.** Starting using pnp/powershell Docker Image might be tricky for beginners **Describe the solution you'd like** An article, similar to https://pnp.github.io/powershell/articles/azurecloudshell.html, that presents different options and different examples of using the module in Docker. **Describe alternatives you've considered** Alternative could be experimenting and learning individually **Additional context**
1.0
Documentation: Using pnp/powershell Docker Image - **Is your feature request related to a problem? Please describe.** Starting using pnp/powershell Docker Image might be tricky for beginners **Describe the solution you'd like** An article, similar to https://pnp.github.io/powershell/articles/azurecloudshell.html, that presents different options and different examples of using the module in Docker. **Describe alternatives you've considered** Alternative could be experimenting and learning individually **Additional context**
non_process
documentation using pnp powershell docker image is your feature request related to a problem please describe starting using pnp powershell docker image might be tricky for beginners describe the solution you d like an article similar to that presents different options and different examples of using the module in docker describe alternatives you ve considered alternative could be experimenting and learning individually additional context
0
14,913
18,296,692,751
IssuesEvent
2021-10-05 21:11:06
pavel-one/Logger
https://api.github.com/repos/pavel-one/Logger
opened
Регистрация и вход ч/з соцсети
Process
Сделать регистрацию через следующие соцсети: 1. Github 2. Google
1.0
Регистрация и вход ч/з соцсети - Сделать регистрацию через следующие соцсети: 1. Github 2. Google
process
регистрация и вход ч з соцсети сделать регистрацию через следующие соцсети github google
1
9,642
12,603,535,545
IssuesEvent
2020-06-11 13:38:21
HackYourFutureBelgium/class-9-10
https://api.github.com/repos/HackYourFutureBelgium/class-9-10
opened
Your Name: module, week
class-10 process-week wednesday-check-in
# Wednesday Check-In __Debugging, Week 1__ ## Progress I learned about JavaScript concepts. ## Blocked So far the process is clear. ## Next Steps Learn more about JavaScript since it requires more practicing to understand it very well. ## Tip(s) of the week I watched videos on YouTube and they were very useful, but they were in Arabic 🤓
1.0
Your Name: module, week - # Wednesday Check-In __Debugging, Week 1__ ## Progress I learned about JavaScript concepts. ## Blocked So far the process is clear. ## Next Steps Learn more about JavaScript since it requires more practicing to understand it very well. ## Tip(s) of the week I watched videos on YouTube and they were very useful, but they were in Arabic 🤓
process
your name module week wednesday check in debugging week progress i learned about javascript concepts blocked so far the process is clear next steps learn more about javascript since it requires more practicing to understand it very well tip s of the week i watched videos on youtube and they were very useful but they were in arabic 🤓
1
17,367
23,191,007,326
IssuesEvent
2022-08-01 12:40:05
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
opened
Release `object_store` `0.4.0`
development-process
As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180 * Planned Release Candidate: 2022-08-08 * Planned Release and Publish to crates.io: 2022-08-11 Items: - [ ] Update changelog and readme: - [ ] Create release scripts - [ ] Create release candidate - [ ] Release candidate approved - [ ] Release to crates.io - [ ] Create release scripts
1.0
Release `object_store` `0.4.0` - As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180 * Planned Release Candidate: 2022-08-08 * Planned Release and Publish to crates.io: 2022-08-11 Items: - [ ] Update changelog and readme: - [ ] Create release scripts - [ ] Create release candidate - [ ] Release candidate approved - [ ] Release to crates.io - [ ] Create release scripts
process
release object store as discussed we will release a new version of object store from the arrow rs repo under asf process planned release candidate planned release and publish to crates io items update changelog and readme create release scripts create release candidate release candidate approved release to crates io create release scripts
1
18,042
24,052,944,827
IssuesEvent
2022-09-16 14:19:03
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Add IERC transferFrom to htsPrecompile approval acceptance tests
enhancement limechain P1 process
### Problem Currently the `htsPrecompile.spec.ts` has tests for approval flow. However, it doesn't follow up with actual transfer tests. ### Solution Add an additional step to each `approval` test to test the `transferFrom()` call. Ensure both `IERC20` and `IERC721` get coverage - `IERC20(token).transferFrom(sender, recipient, amount);` - `IERC721(token).transferFrom(from, to, tokenId);` ### Alternatives _No response_
1.0
Add IERC transferFrom to htsPrecompile approval acceptance tests - ### Problem Currently the `htsPrecompile.spec.ts` has tests for approval flow. However, it doesn't follow up with actual transfer tests. ### Solution Add an additional step to each `approval` test to test the `transferFrom()` call. Ensure both `IERC20` and `IERC721` get coverage - `IERC20(token).transferFrom(sender, recipient, amount);` - `IERC721(token).transferFrom(from, to, tokenId);` ### Alternatives _No response_
process
add ierc transferfrom to htsprecompile approval acceptance tests problem currently the htsprecompile spec ts has tests for approval flow however it doesn t follow up with actual transfer tests solution add an additional step to each approval test to test the transferfrom call ensure both and get coverage token transferfrom sender recipient amount token transferfrom from to tokenid alternatives no response
1
11,884
14,680,446,267
IssuesEvent
2020-12-31 10:05:46
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] [Dev] Studies created from SB having maximum Research Sponsor text are not reflected in Sites/Studies tab in PM
Bug P1 Participant manager Process: Dev Process: Tested QA
Steps: 1. Create a study from SB by giving maximum Research Sponsor text in basic information 2. Launch the study 3. Navigate to PM --> Sites tab 4. Observe the list A/R: Some of the studies created from SB are not reflected in Sites/Studies tab in PM E/R: All studies created from SB should be reflected Study details which is not reflected in PM Study Name: Clinical Trial of the Moventis PNS for Knee Pain Study ID: KneePain App ID: BTCDEV001 Instance: Dev
2.0
[PM] [Dev] Studies created from SB having maximum Research Sponsor text are not reflected in Sites/Studies tab in PM - Steps: 1. Create a study from SB by giving maximum Research Sponsor text in basic information 2. Launch the study 3. Navigate to PM --> Sites tab 4. Observe the list A/R: Some of the studies created from SB are not reflected in Sites/Studies tab in PM E/R: All studies created from SB should be reflected Study details which is not reflected in PM Study Name: Clinical Trial of the Moventis PNS for Knee Pain Study ID: KneePain App ID: BTCDEV001 Instance: Dev
process
studies created from sb having maximum research sponsor text are not reflected in sites studies tab in pm steps create a study from sb by giving maximum research sponsor text in basic information launch the study navigate to pm sites tab observe the list a r some of the studies created from sb are not reflected in sites studies tab in pm e r all studies created from sb should be reflected study details which is not reflected in pm study name clinical trial of the moventis pns for knee pain study id kneepain app id instance dev
1
139,368
18,850,349,884
IssuesEvent
2021-11-11 19:59:18
snowdensb/sonar-xanitizer
https://api.github.com/repos/snowdensb/sonar-xanitizer
opened
CVE-2020-24616 (High) detected in jackson-databind-2.6.3.jar
security vulnerability
## CVE-2020-24616 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /src/test/resources/webgoat/WEB-INF/lib/jackson-databind-2.6.3.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowdensb/sonar-xanitizer/commit/e2144e84b1fdbf18c01c24e6ab9ade7b45b25283">e2144e84b1fdbf18c01c24e6ab9ade7b45b25283</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP). <p>Publish Date: 2020-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-25</p> <p>Fix Resolution: 2.9.10.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.3","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-24616 (High) detected in jackson-databind-2.6.3.jar - ## CVE-2020-24616 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /src/test/resources/webgoat/WEB-INF/lib/jackson-databind-2.6.3.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowdensb/sonar-xanitizer/commit/e2144e84b1fdbf18c01c24e6ab9ade7b45b25283">e2144e84b1fdbf18c01c24e6ab9ade7b45b25283</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP). <p>Publish Date: 2020-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-25</p> <p>Fix Resolution: 2.9.10.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.3","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library src test resources webgoat web inf lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp vulnerabilityurl
0
3,728
6,733,142,338
IssuesEvent
2017-10-18 13:58:37
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
closed
Contract Preparation Detail Wizard
form process workflow
Two Steps Wizard Wireframe ![image](https://user-images.githubusercontent.com/3499016/30181609-d36070de-93e1-11e7-8353-ec0d2c72fb04.png) Dynamically load fields that are applicable for the chosen activity type. **After submitted, only update in the collection**
1.0
Contract Preparation Detail Wizard - Two Steps Wizard Wireframe ![image](https://user-images.githubusercontent.com/3499016/30181609-d36070de-93e1-11e7-8353-ec0d2c72fb04.png) Dynamically load fields that are applicable for the chosen activity type. **After submitted, only update in the collection**
process
contract preparation detail wizard two steps wizard wireframe dynamically load fields that are applicable for the chosen activity type after submitted only update in the collection
1
614,226
19,161,598,265
IssuesEvent
2021-12-03 01:15:45
returntocorp/semgrep
https://api.github.com/repos/returntocorp/semgrep
closed
Opposite conditional matching behavior in PHP and Java
priority:low lang:java lang:php feature:matching stale
In Java matching conditionals with ``` if (...) { ... } ``` matches an if-block as well as an if-else-block. In PHP only the if-block is matched. However, changing the pattern to ``` if (...) { ... } else { ... } ``` will have the opposite behavior in either languages. Links: https://semgrep.dev/s/bashprince:java https://semgrep.dev/s/bashprince:php Not sure which one is the correct behavior, but I would think the Java version makes more sense in this case since it matches the whole block and the further specification with else makes the match more exact. - [ ] P2: regular bug that should get fixed Tested on semgrep.dev
1.0
Opposite conditional matching behavior in PHP and Java - In Java matching conditionals with ``` if (...) { ... } ``` matches an if-block as well as an if-else-block. In PHP only the if-block is matched. However, changing the pattern to ``` if (...) { ... } else { ... } ``` will have the opposite behavior in either languages. Links: https://semgrep.dev/s/bashprince:java https://semgrep.dev/s/bashprince:php Not sure which one is the correct behavior, but I would think the Java version makes more sense in this case since it matches the whole block and the further specification with else makes the match more exact. - [ ] P2: regular bug that should get fixed Tested on semgrep.dev
non_process
opposite conditional matching behavior in php and java in java matching conditionals with if matches an if block as well as an if else block in php only the if block is matched however changing the pattern to if else will have the opposite behavior in either languages links not sure which one is the correct behavior but i would think the java version makes more sense in this case since it matches the whole block and the further specification with else makes the match more exact regular bug that should get fixed tested on semgrep dev
0
121,701
16,016,616,584
IssuesEvent
2021-04-20 16:48:04
microsoft/BotFramework-Composer
https://api.github.com/repos/microsoft/BotFramework-Composer
closed
Restart selected bot with single click
Bot Services UX Design customer-reported feature-request
## Is your feature request related to a problem? Please describe. #7049 If multiple bot is running , there is no easy solution to restart the particular bot , current approach we have restart all the bots or stop and start selected bot. ![image](https://user-images.githubusercontent.com/16264167/115144257-02ced480-a04c-11eb-9e3d-76c5df3c7fe0.png) ## Describe the solution you'd like If user has selected overall project ![image](https://user-images.githubusercontent.com/16264167/115144402-b8018c80-a04c-11eb-922c-5d4ca8f20ce6.png) If user selected particular bot then restart only particular bot instead of all ![image](https://user-images.githubusercontent.com/16264167/115144420-cb145c80-a04c-11eb-982c-35a40d7ccff6.png) ![image](https://user-images.githubusercontent.com/16264167/115144512-58f04780-a04d-11eb-9cc1-8356228f156d.png)
1.0
Restart selected bot with single click - ## Is your feature request related to a problem? Please describe. #7049 If multiple bot is running , there is no easy solution to restart the particular bot , current approach we have restart all the bots or stop and start selected bot. ![image](https://user-images.githubusercontent.com/16264167/115144257-02ced480-a04c-11eb-9e3d-76c5df3c7fe0.png) ## Describe the solution you'd like If user has selected overall project ![image](https://user-images.githubusercontent.com/16264167/115144402-b8018c80-a04c-11eb-922c-5d4ca8f20ce6.png) If user selected particular bot then restart only particular bot instead of all ![image](https://user-images.githubusercontent.com/16264167/115144420-cb145c80-a04c-11eb-982c-35a40d7ccff6.png) ![image](https://user-images.githubusercontent.com/16264167/115144512-58f04780-a04d-11eb-9cc1-8356228f156d.png)
non_process
restart selected bot with single click is your feature request related to a problem please describe if multiple bot is running there is no easy solution to restart the particular bot current approach we have restart all the bots or stop and start selected bot describe the solution you d like if user has selected overall project if user selected particular bot then restart only particular bot instead of all
0
434,577
30,456,278,918
IssuesEvent
2023-07-16 23:12:32
CarsOk/droguerias_darcy
https://api.github.com/repos/CarsOk/droguerias_darcy
closed
Diagrama entidad relacion
documentation
## Yo como Aprendiz SENA ## Quiero Subir el diagrama entidad relacion del proyecto ## Para Seguimiento inicial
1.0
Diagrama entidad relacion - ## Yo como Aprendiz SENA ## Quiero Subir el diagrama entidad relacion del proyecto ## Para Seguimiento inicial
non_process
diagrama entidad relacion yo como aprendiz sena quiero subir el diagrama entidad relacion del proyecto para seguimiento inicial
0
87,711
25,192,032,325
IssuesEvent
2022-11-12 03:28:19
tensorflow/tfjs
https://api.github.com/repos/tensorflow/tfjs
closed
Compile error with @tensorflow/tfjs-backend-webgl for TS target > ES5
type:build/install
**System information** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: - - TensorFlow.js installed from (npm or script link): npm - TensorFlow.js version: 2.7.0 - CUDA/cuDNN version: - **Describe the problem** When trying to build a typescript code that depends on `@tensorflow/tfjs-backend-webgl`, if the `compilerOptions.target` is above ES5 (i.e "ES2015","ES2016","ES2017","ES2018","ES2019","ES2020" or "ESNext"), you get the following compilation errors: ``` node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextBase'. Types of property 'clearBufferfv' are incompatible. Type '(buffer: number, drawbuffer: number, values: ArrayLike<number> | Float32Array, srcOffset?: number) => void' is not assignable to type '{ (buffer: number, drawbuffer: number, values: Float32List, srcOffset?: number): void; (buffer: number, drawbuffer: number, values: Iterable<number>, srcOffset?: number): void; }'. Types of parameters 'values' and 'values' are incompatible. Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'. Type 'Iterable<number>' is missing the following properties from type 'Float32Array': BYTES_PER_ELEMENT, buffer, byteLength, byteOffset, and 25 more. 16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase { ~~~~~~~~~~~~~~~~~~~~~~ node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextOverloads'. Types of property 'uniform1fv' are incompatible. Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, data: Float32List, srcOffset?: number, srcLength?: number): void; (location: WebGLUniformLocation, data: Iterable<...>, srcOffset?: number, srcLength?: number): void; }'. Types of parameters 'data' and 'data' are incompatible. Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'. Type 'Iterable<number>' is not assignable to type 'Float32Array'. 16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase { ~~~~~~~~~~~~~~~~~~~~~~ node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGLRenderingContext'. Types of property 'uniform1fv' are incompatible. Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, v: Float32List): void; (location: WebGLUniformLocation, v: Iterable<number>): void; }'. Types of parameters 'data' and 'v' are incompatible. Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'. Type 'Iterable<number>' is not assignable to type 'Float32Array'. 16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase { ~~~~~~~~~~~~~~~~~~~~~~ Found 3 errors. ``` **Provide the exact sequence of commands / steps that you executed before running into the problem** The error is easily reconstructed with the following: 1. package.json: ``` { "name": "tfjs-ts-test", "version": "1.0.0", "description": "", "main": "index.js", "scripts": { "build": "tsc" }, "author": "", "license": "ISC", "dependencies": { "@tensorflow/tfjs-backend-webgl": "^2.7.0", "@tensorflow/tfjs-core": "^2.7.0", "typescript": "^4.0.5" } } ``` 2. tsconfig.json: ``` { "compilerOptions": { "target": "ES2015" } } ``` 3. An empty index.ts 4. npm install 5. npm run build **Any other info / logs** This partially relates to https://github.com/tensorflow/tfjs/issues/2007 but not the same issue
1.0
Compile error with @tensorflow/tfjs-backend-webgl for TS target > ES5 - **System information** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: - - TensorFlow.js installed from (npm or script link): npm - TensorFlow.js version: 2.7.0 - CUDA/cuDNN version: - **Describe the problem** When trying to build a typescript code that depends on `@tensorflow/tfjs-backend-webgl`, if the `compilerOptions.target` is above ES5 (i.e "ES2015","ES2016","ES2017","ES2018","ES2019","ES2020" or "ESNext"), you get the following compilation errors: ``` node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextBase'. Types of property 'clearBufferfv' are incompatible. Type '(buffer: number, drawbuffer: number, values: ArrayLike<number> | Float32Array, srcOffset?: number) => void' is not assignable to type '{ (buffer: number, drawbuffer: number, values: Float32List, srcOffset?: number): void; (buffer: number, drawbuffer: number, values: Iterable<number>, srcOffset?: number): void; }'. Types of parameters 'values' and 'values' are incompatible. Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'. Type 'Iterable<number>' is missing the following properties from type 'Float32Array': BYTES_PER_ELEMENT, buffer, byteLength, byteOffset, and 25 more. 16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase { ~~~~~~~~~~~~~~~~~~~~~~ node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextOverloads'. Types of property 'uniform1fv' are incompatible. Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, data: Float32List, srcOffset?: number, srcLength?: number): void; (location: WebGLUniformLocation, data: Iterable<...>, srcOffset?: number, srcLength?: number): void; }'. Types of parameters 'data' and 'data' are incompatible. Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'. Type 'Iterable<number>' is not assignable to type 'Float32Array'. 16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase { ~~~~~~~~~~~~~~~~~~~~~~ node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGLRenderingContext'. Types of property 'uniform1fv' are incompatible. Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, v: Float32List): void; (location: WebGLUniformLocation, v: Iterable<number>): void; }'. Types of parameters 'data' and 'v' are incompatible. Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'. Type 'Iterable<number>' is not assignable to type 'Float32Array'. 16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase { ~~~~~~~~~~~~~~~~~~~~~~ Found 3 errors. ``` **Provide the exact sequence of commands / steps that you executed before running into the problem** The error is easily reconstructed with the following: 1. package.json: ``` { "name": "tfjs-ts-test", "version": "1.0.0", "description": "", "main": "index.js", "scripts": { "build": "tsc" }, "author": "", "license": "ISC", "dependencies": { "@tensorflow/tfjs-backend-webgl": "^2.7.0", "@tensorflow/tfjs-core": "^2.7.0", "typescript": "^4.0.5" } } ``` 2. tsconfig.json: ``` { "compilerOptions": { "target": "ES2015" } } ``` 3. An empty index.ts 4. npm install 5. npm run build **Any other info / logs** This partially relates to https://github.com/tensorflow/tfjs/issues/2007 but not the same issue
non_process
compile error with tensorflow tfjs backend webgl for ts target system information os platform and distribution e g linux ubuntu windows mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow js installed from npm or script link npm tensorflow js version cuda cudnn version describe the problem when trying to build a typescript code that depends on tensorflow tfjs backend webgl if the compileroptions target is above i e or esnext you get the following compilation errors node modules typescript lib lib dom d ts error interface incorrectly extends interface types of property clearbufferfv are incompatible type buffer number drawbuffer number values arraylike srcoffset number void is not assignable to type buffer number drawbuffer number values srcoffset number void buffer number drawbuffer number values iterable srcoffset number void types of parameters values and values are incompatible type iterable is not assignable to type arraylike type iterable is missing the following properties from type bytes per element buffer bytelength byteoffset and more interface extends webglrenderingcontextbase node modules typescript lib lib dom d ts error interface incorrectly extends interface types of property are incompatible type location webgluniformlocation data arraylike srcoffset number srclength number void is not assignable to type location webgluniformlocation data srcoffset number srclength number void location webgluniformlocation data iterable srcoffset number srclength number void types of parameters data and data are incompatible type iterable is not assignable to type arraylike type iterable is not assignable to type interface extends webglrenderingcontextbase node modules typescript lib lib dom d ts error interface incorrectly extends interface webglrenderingcontext types of property are incompatible type location webgluniformlocation data arraylike srcoffset number srclength number void is not assignable to type location webgluniformlocation v void location webgluniformlocation v iterable void types of parameters data and v are incompatible type iterable is not assignable to type arraylike type iterable is not assignable to type interface extends webglrenderingcontextbase found errors provide the exact sequence of commands steps that you executed before running into the problem the error is easily reconstructed with the following package json name tfjs ts test version description main index js scripts build tsc author license isc dependencies tensorflow tfjs backend webgl tensorflow tfjs core typescript tsconfig json compileroptions target an empty index ts npm install npm run build any other info logs this partially relates to but not the same issue
0
258,224
27,563,872,095
IssuesEvent
2023-03-08 01:12:29
billmcchesney1/t-vault
https://api.github.com/repos/billmcchesney1/t-vault
opened
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz
security vulnerability
## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: /tvaultui/package.json</p> <p>Path to vulnerable library: /tvaultui/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - gulp-sass-3.1.0.tgz (Root Library) - node-sass-4.14.1.tgz - meow-3.7.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution (hosted-git-info): 2.8.9</p> <p>Direct dependency fix Resolution (gulp-sass): 3.2.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: /tvaultui/package.json</p> <p>Path to vulnerable library: /tvaultui/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - gulp-sass-3.1.0.tgz (Root Library) - node-sass-4.14.1.tgz - meow-3.7.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution (hosted-git-info): 2.8.9</p> <p>Direct dependency fix Resolution (gulp-sass): 3.2.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_process
cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file tvaultui package json path to vulnerable library tvaultui node modules hosted git info package json dependency hierarchy gulp sass tgz root library node sass tgz meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in base branch dev vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info direct dependency fix resolution gulp sass check this box to open an automated fix pr
0
40,869
10,586,892,721
IssuesEvent
2019-10-08 20:45:33
ged/ruby-pg
https://api.github.com/repos/ged/ruby-pg
closed
gem install pg on windows 7 fails
0.10.1 Build System bug major
**[Original report](https://bitbucket.org/ged/ruby-pg/issue/64) by Anonymous.** ---------------------------------------- When I run "bundle install", I get this error: checking for pg_config... yes Using config values from c:\PostgreSQL\9.0\bin/pg_config.exe checking for libpq-fe.h... yes checking for libpq/libpq-fs.h... yes checking for PQconnectdb() in -lpq... no checking for PQconnectdb() in -llibpq... no checking for PQconnectdb() in -lms/libpq... no Can't find the PostgreSQL client library (libpq) *** extconf.rb failed *** Could not create Makefile due to some reason, probably lack of necessary libraries and/or headers. Check the mkmf.log file for more details. You may need configuration options.
1.0
gem install pg on windows 7 fails - **[Original report](https://bitbucket.org/ged/ruby-pg/issue/64) by Anonymous.** ---------------------------------------- When I run "bundle install", I get this error: checking for pg_config... yes Using config values from c:\PostgreSQL\9.0\bin/pg_config.exe checking for libpq-fe.h... yes checking for libpq/libpq-fs.h... yes checking for PQconnectdb() in -lpq... no checking for PQconnectdb() in -llibpq... no checking for PQconnectdb() in -lms/libpq... no Can't find the PostgreSQL client library (libpq) *** extconf.rb failed *** Could not create Makefile due to some reason, probably lack of necessary libraries and/or headers. Check the mkmf.log file for more details. You may need configuration options.
non_process
gem install pg on windows fails by anonymous when i run bundle install i get this error checking for pg config yes using config values from c postgresql bin pg config exe checking for libpq fe h yes checking for libpq libpq fs h yes checking for pqconnectdb in lpq no checking for pqconnectdb in llibpq no checking for pqconnectdb in lms libpq no can t find the postgresql client library libpq extconf rb failed could not create makefile due to some reason probably lack of necessary libraries and or headers check the mkmf log file for more details you may need configuration options
0
184,588
14,289,501,639
IssuesEvent
2020-11-23 19:21:45
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
zhanshicai/zhanshicai: gossip/gossip/orgs_test.go; 14 LoC
fresh small test
Found a possible issue in [zhanshicai/zhanshicai](https://www.github.com/zhanshicai/zhanshicai) at [gossip/gossip/orgs_test.go](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for _, ch := range channels { if isOrgInChan(org, ch) { for _, p := range peers { p.JoinChan(joinChanMsgsByChan[ch], common.ChainID(ch)) p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch)) go func(p Gossip) { for i := 0; i < 5; i++ { time.Sleep(time.Second) p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch)) } }(p) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096
1.0
zhanshicai/zhanshicai: gossip/gossip/orgs_test.go; 14 LoC - Found a possible issue in [zhanshicai/zhanshicai](https://www.github.com/zhanshicai/zhanshicai) at [gossip/gossip/orgs_test.go](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for _, ch := range channels { if isOrgInChan(org, ch) { for _, p := range peers { p.JoinChan(joinChanMsgsByChan[ch], common.ChainID(ch)) p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch)) go func(p Gossip) { for i := 0; i < 5; i++ { time.Sleep(time.Second) p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch)) } }(p) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096
non_process
zhanshicai zhanshicai gossip gossip orgs test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for ch range channels if isorginchan org ch for p range peers p joinchan joinchanmsgsbychan common chainid ch p updatechannelmetadata createmetadata common chainid ch go func p gossip for i i i time sleep time second p updatechannelmetadata createmetadata common chainid ch p leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
359,175
25,224,407,260
IssuesEvent
2022-11-14 15:01:57
stratosphereips/nist-cve-search-tool
https://api.github.com/repos/stratosphereips/nist-cve-search-tool
closed
Add documentation on how to use the Docker image
documentation
There's now a docker image, mostly for ensuring LTS. Add to the readme that the docker image exists, how to get it and a few examples of how to use it.
1.0
Add documentation on how to use the Docker image - There's now a docker image, mostly for ensuring LTS. Add to the readme that the docker image exists, how to get it and a few examples of how to use it.
non_process
add documentation on how to use the docker image there s now a docker image mostly for ensuring lts add to the readme that the docker image exists how to get it and a few examples of how to use it
0
452,730
32,066,489,936
IssuesEvent
2023-09-25 03:45:46
apache/incubator-opendal
https://api.github.com/repos/apache/incubator-opendal
closed
docs: Update the announcement email template to ensure that disclaimers have been added
documentation good first issue help wanted
As mentioned in the discussion at https://lists.apache.org/thread/0d8dxxbhb28m7w7kqxs1d2n0hsdybocg, Sebb reminded us to include disclaimers in our announcement. We should add the following content: ``` --- Apache OpenDAL (incubating) is an effort undergoing incubation at the Apache Software Foundation (ASF), sponsored by the Apache Incubator PMC. Incubation is required of all newly accepted projects until a further review indicates that the infrastructure, communications, and decision making process have stabilized in a manner consistent with other successful ASF projects. While incubation status is not necessarily a reflection of the completeness or stability of the code, it does indicate that the project has yet to be fully endorsed by the ASF. ```
1.0
docs: Update the announcement email template to ensure that disclaimers have been added - As mentioned in the discussion at https://lists.apache.org/thread/0d8dxxbhb28m7w7kqxs1d2n0hsdybocg, Sebb reminded us to include disclaimers in our announcement. We should add the following content: ``` --- Apache OpenDAL (incubating) is an effort undergoing incubation at the Apache Software Foundation (ASF), sponsored by the Apache Incubator PMC. Incubation is required of all newly accepted projects until a further review indicates that the infrastructure, communications, and decision making process have stabilized in a manner consistent with other successful ASF projects. While incubation status is not necessarily a reflection of the completeness or stability of the code, it does indicate that the project has yet to be fully endorsed by the ASF. ```
non_process
docs update the announcement email template to ensure that disclaimers have been added as mentioned in the discussion at sebb reminded us to include disclaimers in our announcement we should add the following content apache opendal incubating is an effort undergoing incubation at the apache software foundation asf sponsored by the apache incubator pmc incubation is required of all newly accepted projects until a further review indicates that the infrastructure communications and decision making process have stabilized in a manner consistent with other successful asf projects while incubation status is not necessarily a reflection of the completeness or stability of the code it does indicate that the project has yet to be fully endorsed by the asf
0
21,315
11,188,430,039
IssuesEvent
2020-01-02 04:57:52
0xProject/OpenZKP
https://api.github.com/repos/0xProject/OpenZKP
closed
Special case z == FieldElement::ONE?
performance tracker
*On 2019-04-23 @Recmo wrote in [`87f22ab`](https://github.com/0xProject/OpenZKP/commit/87f22ab866dbb5241a13c5916a726dd6047ed33d) “Implement edge cases in Jacobian”:* Special case z == FieldElement::ONE? See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl ```rust self.x = x.clone(); self.y = y.clone(); self.z = FieldElement::ONE; return; } // OPT: Special case z == FieldElement::ONE? // See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl let z1z1 = self.z.square(); let u2 = x * &z1z1; let s2 = y * &self.z * &z1z1; if self.x == u2 { return if self.x == s2 { ``` *From [`algebra/elliptic-curve/src/jacobian.rs:199`](https://github.com/0xProject/OpenZKP/blob/3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2/algebra/elliptic-curve/src/jacobian.rs#L199)* <!--{"commit-hash": "87f22ab866dbb5241a13c5916a726dd6047ed33d", "author": "Remco Bloemen", "author-mail": "<remco@0x.org>", "author-time": 1556062052, "author-tz": "-0700", "committer": "Remco Bloemen", "committer-mail": "<remco@0x.org>", "committer-time": 1556062052, "committer-tz": "-0700", "summary": "Implement edge cases in Jacobian", "previous": "77dc5773867b7ce66823b331137a1e64b353eb1c src/jacobian.rs", "filename": "algebra/elliptic-curve/src/jacobian.rs", "line": 198, "line_end": 200, "kind": "OPT", "issue": "Special case z == FieldElement::ONE?\nSee http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl", "head": "Special case z == FieldElement::ONE?", "context": " self.x = x.clone();\n self.y = y.clone();\n self.z = FieldElement::ONE;\n return;\n }\n // OPT: Special case z == FieldElement::ONE?\n // See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl\n let z1z1 = self.z.square();\n let u2 = x * &z1z1;\n let s2 = y * &self.z * &z1z1;\n if self.x == u2 {\n return if self.x == s2 {\n", "repo": "0xProject/OpenZKP", "branch-hash": "3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2"}-->
True
Special case z == FieldElement::ONE? - *On 2019-04-23 @Recmo wrote in [`87f22ab`](https://github.com/0xProject/OpenZKP/commit/87f22ab866dbb5241a13c5916a726dd6047ed33d) “Implement edge cases in Jacobian”:* Special case z == FieldElement::ONE? See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl ```rust self.x = x.clone(); self.y = y.clone(); self.z = FieldElement::ONE; return; } // OPT: Special case z == FieldElement::ONE? // See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl let z1z1 = self.z.square(); let u2 = x * &z1z1; let s2 = y * &self.z * &z1z1; if self.x == u2 { return if self.x == s2 { ``` *From [`algebra/elliptic-curve/src/jacobian.rs:199`](https://github.com/0xProject/OpenZKP/blob/3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2/algebra/elliptic-curve/src/jacobian.rs#L199)* <!--{"commit-hash": "87f22ab866dbb5241a13c5916a726dd6047ed33d", "author": "Remco Bloemen", "author-mail": "<remco@0x.org>", "author-time": 1556062052, "author-tz": "-0700", "committer": "Remco Bloemen", "committer-mail": "<remco@0x.org>", "committer-time": 1556062052, "committer-tz": "-0700", "summary": "Implement edge cases in Jacobian", "previous": "77dc5773867b7ce66823b331137a1e64b353eb1c src/jacobian.rs", "filename": "algebra/elliptic-curve/src/jacobian.rs", "line": 198, "line_end": 200, "kind": "OPT", "issue": "Special case z == FieldElement::ONE?\nSee http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl", "head": "Special case z == FieldElement::ONE?", "context": " self.x = x.clone();\n self.y = y.clone();\n self.z = FieldElement::ONE;\n return;\n }\n // OPT: Special case z == FieldElement::ONE?\n // See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl\n let z1z1 = self.z.square();\n let u2 = x * &z1z1;\n let s2 = y * &self.z * &z1z1;\n if self.x == u2 {\n return if self.x == s2 {\n", "repo": "0xProject/OpenZKP", "branch-hash": "3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2"}-->
non_process
special case z fieldelement one on recmo wrote in “implement edge cases in jacobian” special case z fieldelement one see rust self x x clone self y y clone self z fieldelement one return opt special case z fieldelement one see let self z square let x let y self z if self x return if self x from author time author tz committer remco bloemen committer mail committer time committer tz summary implement edge cases in jacobian previous src jacobian rs filename algebra elliptic curve src jacobian rs line line end kind opt issue special case z fieldelement one nsee head special case z fieldelement one context self x x clone n self y y clone n self z fieldelement one n return n n opt special case z fieldelement one n see let self z square n let x n let y self z n if self x n return if self x n repo openzkp branch hash
0
1,880
4,019,362,958
IssuesEvent
2016-05-16 14:41:41
psu-libraries/library_services_model
https://api.github.com/repos/psu-libraries/library_services_model
opened
Join everyone to the repo
services development
### Requirements - [ ] Rama invited - [ ] Stephen invited - [ ] Patricia invited - [ ] Nathan invited - [ ] Robyn invited
1.0
Join everyone to the repo - ### Requirements - [ ] Rama invited - [ ] Stephen invited - [ ] Patricia invited - [ ] Nathan invited - [ ] Robyn invited
non_process
join everyone to the repo requirements rama invited stephen invited patricia invited nathan invited robyn invited
0
12,357
14,887,185,073
IssuesEvent
2021-01-20 17:58:12
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Open study > Enrollment target > Entered data is getting removed
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
AR : Open study > Enrollment target > Entered data is getting removed when user enters the 6th digit ER : It should not allow the user to enter 6th digit and entered data should not be removed [ Note : It should also be fixed when user enters other characters after numbers ] https://user-images.githubusercontent.com/60386291/103529284-65400280-4eab-11eb-82e6-015be0401bd1.mp4
3.0
[PM] Open study > Enrollment target > Entered data is getting removed - AR : Open study > Enrollment target > Entered data is getting removed when user enters the 6th digit ER : It should not allow the user to enter 6th digit and entered data should not be removed [ Note : It should also be fixed when user enters other characters after numbers ] https://user-images.githubusercontent.com/60386291/103529284-65400280-4eab-11eb-82e6-015be0401bd1.mp4
process
open study enrollment target entered data is getting removed ar open study enrollment target entered data is getting removed when user enters the digit er it should not allow the user to enter digit and entered data should not be removed
1
271,851
20,719,592,014
IssuesEvent
2022-03-13 06:52:02
christian-cahig/Masterarbeit-APF
https://api.github.com/repos/christian-cahig/Masterarbeit-APF
opened
Better naming and notation for "generalized branch connection matrix" and "intermediate state vector"
documentation enhancement PyAPF APF.m
To be consistent with the "full" and "reduced" versions of some vectors and matrices, there should be a _full_ and a _reduced intermediate state vectors_ denoted by `$ \boldsymbol{x} $` (`x`) and `$ \boldsymbol{w} $` (`w`), respectively. Moreover, it is a bit misleading to use the term "generalized branch connection matrix" when all that has been done is augment the directed and undirected branch-to-bus incidence matrices. It seems more accurate to therefore use _augmented branch-to-bus incidence matrix_. This will come in two variants: _full_ (`$ \boldsymbol{C} $` and `C`) and _reduced_ (`$ \boldsymbol{U} $` and `U`). Hence, the factorized formulation of the power flow equations will still consist of `$ \boldsymbol{E} \boldsymbol{y} = \boldsymbol{e} $` and `$ \boldsymbol{C} \boldsymbol{x} = \boldsymbol{u} $`, but the ones used in solving for the anticipated voltages are `$ \boldsymbol{Z} \boldsymbol{y} = \boldsymbol{z} $` and `$ \boldsymbol{U} \boldsymbol{w} = \boldsymbol{u} $`. Consequently, the matrix currently denoted by `$ \boldsymbol{U} $` (`U`) needs to be assigned another symbol; `$ \boldsymbol{F} $` (`F`) seems fine.
1.0
Better naming and notation for "generalized branch connection matrix" and "intermediate state vector" - To be consistent with the "full" and "reduced" versions of some vectors and matrices, there should be a _full_ and a _reduced intermediate state vectors_ denoted by `$ \boldsymbol{x} $` (`x`) and `$ \boldsymbol{w} $` (`w`), respectively. Moreover, it is a bit misleading to use the term "generalized branch connection matrix" when all that has been done is augment the directed and undirected branch-to-bus incidence matrices. It seems more accurate to therefore use _augmented branch-to-bus incidence matrix_. This will come in two variants: _full_ (`$ \boldsymbol{C} $` and `C`) and _reduced_ (`$ \boldsymbol{U} $` and `U`). Hence, the factorized formulation of the power flow equations will still consist of `$ \boldsymbol{E} \boldsymbol{y} = \boldsymbol{e} $` and `$ \boldsymbol{C} \boldsymbol{x} = \boldsymbol{u} $`, but the ones used in solving for the anticipated voltages are `$ \boldsymbol{Z} \boldsymbol{y} = \boldsymbol{z} $` and `$ \boldsymbol{U} \boldsymbol{w} = \boldsymbol{u} $`. Consequently, the matrix currently denoted by `$ \boldsymbol{U} $` (`U`) needs to be assigned another symbol; `$ \boldsymbol{F} $` (`F`) seems fine.
non_process
better naming and notation for generalized branch connection matrix and intermediate state vector to be consistent with the full and reduced versions of some vectors and matrices there should be a full and a reduced intermediate state vectors denoted by boldsymbol x x and boldsymbol w w respectively moreover it is a bit misleading to use the term generalized branch connection matrix when all that has been done is augment the directed and undirected branch to bus incidence matrices it seems more accurate to therefore use augmented branch to bus incidence matrix this will come in two variants full boldsymbol c and c and reduced boldsymbol u and u hence the factorized formulation of the power flow equations will still consist of boldsymbol e boldsymbol y boldsymbol e and boldsymbol c boldsymbol x boldsymbol u but the ones used in solving for the anticipated voltages are boldsymbol z boldsymbol y boldsymbol z and boldsymbol u boldsymbol w boldsymbol u consequently the matrix currently denoted by boldsymbol u u needs to be assigned another symbol boldsymbol f f seems fine
0
22,146
7,124,795,382
IssuesEvent
2018-01-19 20:16:27
linode/manager
https://api.github.com/repos/linode/manager
closed
Upgrade to React 16
Backlog Build & Organization
This builds off of #2676 React 16 provides fibers, which should offer performance updates to our app, error boundaries, and a smaller footprint: - https://edgecoders.com/react-16-features-and-fiber-explanation-e779544bb1b7 - https://reactjs.org/blog/2017/09/26/react-v16.0.html - https://medium.com/netscape/whats-new-in-react-16-1608390ffe39 I took a stab at this but found that some of our external components have not been updated for React 16: - https://github.com/akiran/react-highlight/pull/43 - https://github.com/HubSpot/react-select-plus/issues/94 - a few others We should pursue those PRs (some are very stale) or replace those components with more active alternatives.
1.0
Upgrade to React 16 - This builds off of #2676 React 16 provides fibers, which should offer performance updates to our app, error boundaries, and a smaller footprint: - https://edgecoders.com/react-16-features-and-fiber-explanation-e779544bb1b7 - https://reactjs.org/blog/2017/09/26/react-v16.0.html - https://medium.com/netscape/whats-new-in-react-16-1608390ffe39 I took a stab at this but found that some of our external components have not been updated for React 16: - https://github.com/akiran/react-highlight/pull/43 - https://github.com/HubSpot/react-select-plus/issues/94 - a few others We should pursue those PRs (some are very stale) or replace those components with more active alternatives.
non_process
upgrade to react this builds off of react provides fibers which should offer performance updates to our app error boundaries and a smaller footprint i took a stab at this but found that some of our external components have not been updated for react a few others we should pursue those prs some are very stale or replace those components with more active alternatives
0
22,169
30,719,861,849
IssuesEvent
2023-07-27 15:13:22
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
@saltcorn/mobile-builder 0.8.7 has 3 guarddog issues
npm-install-script npm-silent-process-execution
```{"npm-install-script":[{"code":" \"postinstall\": \"node ./docker/post-installer.js\"","location":"package/package.json:10","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/dist/docker/post-installer.js:25","message":"This package is silently executing another executable"},{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/docker/post-installer.js:22","message":"This package is silently executing another executable"}]}```
1.0
@saltcorn/mobile-builder 0.8.7 has 3 guarddog issues - ```{"npm-install-script":[{"code":" \"postinstall\": \"node ./docker/post-installer.js\"","location":"package/package.json:10","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/dist/docker/post-installer.js:25","message":"This package is silently executing another executable"},{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/docker/post-installer.js:22","message":"This package is silently executing another executable"}]}```
process
saltcorn mobile builder has guarddog issues npm install script npm silent process execution
1
17,847
23,785,049,575
IssuesEvent
2022-09-02 09:18:09
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Improve logs in basic error reports
process/candidate kind/improvement tech/typescript topic: error reporting team/client
In many crash reports the logs look like this: ``` prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true ``` Example: https://github.com/prisma/prisma/issues/14248 We want to have more useful information here and probably don't need to log `libraryStarted` each time.
1.0
Improve logs in basic error reports - In many crash reports the logs look like this: ``` prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true prisma:client:libraryEngine sending request, this.libraryStarted: true ``` Example: https://github.com/prisma/prisma/issues/14248 We want to have more useful information here and probably don't need to log `libraryStarted` each time.
process
improve logs in basic error reports in many crash reports the logs look like this prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true example we want to have more useful information here and probably don t need to log librarystarted each time
1
5,371
8,202,228,676
IssuesEvent
2018-09-02 06:12:52
bio-miga/miga
https://api.github.com/repos/bio-miga/miga
closed
Re-registration of previous results not working
API Processing bug
Some steps involve (un)zipping files from previous steps, and registering the result should trigger a re-registration. However, the presence of the previous .json file stops this from happening. Also, this re-registration should be moved from the `miga-base` Ruby code to the execution Bash scripts. It may need a new `--force` option for `miga add_result`.
1.0
Re-registration of previous results not working - Some steps involve (un)zipping files from previous steps, and registering the result should trigger a re-registration. However, the presence of the previous .json file stops this from happening. Also, this re-registration should be moved from the `miga-base` Ruby code to the execution Bash scripts. It may need a new `--force` option for `miga add_result`.
process
re registration of previous results not working some steps involve un zipping files from previous steps and registering the result should trigger a re registration however the presence of the previous json file stops this from happening also this re registration should be moved from the miga base ruby code to the execution bash scripts it may need a new force option for miga add result
1
21,406
11,660,229,153
IssuesEvent
2020-03-03 02:37:44
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
opened
Feature: Unpaid Fee Flag on Permit Issuance
Product: AMANDA Project: ATD AMANDA Backlog Service: Apps Type: Enhancement Workgroup: ROW migrated
## Unpaid Fee Flag _AKA # 46 Unpaid Fee Flag on Permit Issuance (Audit Request)_ **Stakeholders:** Paloma, Kim, Mooney **User Story:** As a ROW permit analyst (PLA), I need to know if any of the folder people have unpaid ROW fees BEFORE the permit is issued, so I can place the permit on hold and attempt to collect the outstanding fee(s). **Current State:** 1. The PLA, prior to permit approval, performs manual research to identify any unpaid ROW fees for ALL people listed on the people tab (Contractor, applicant, contact, whoever). - They do not collect fees billed before 10/1/2013. 2. If no outstanding fees are identified, the permit is issued. End process. 3. If outstanding ATD fees are identified, the PLA provides the outstanding fee information to the person/entity responsible for payment. - So if there are 3 unique people/entities on the permit, each entity will be provided with their unique fee information. - The people/entities are provided with instructions for paying, all outstanding bill numbers, and the total for each bill number. 4. The PLA adds a note to the folder description: "Pending Outstanding Fee Payment **TL" 5. The customer will pay the fees, either online or in person. 6. The customer notifies PLA of payment - When this step is missed, it leads to customers working in the ROW without an approved permit 7. PLA issues the permit. End Process. **Future State:** 1. Folder issuance is attempted, either by a user from the folder tab, or by a system procedure like post payment. - We need to be sure any new procedure doesn't cause payment update failures when the fees are paid online. (Put simply, lets learn from the issues caused by the Post Payment upgrade.) - This feature may have to be split into 2 features, one for running this procedure when issuance is initiated by the user from the folder tab, and another for running this procedure when issuance is initiated by a post payment procedure. **Will need LIT feedback on this.** 2. The system performs the unpaid fee verification - See # 1 in Current State; the system needs to do this instead of the PLA - Query all unpaid fees for each people record on the people tab. Of those fees, exclude any not in folder types RW, EX, LC, ECV, GF, or DS. Of those fees, exclude any created prior to 10/1/2013. **Now you have a list of qualifying fees.** 3. If **qualifying fees** are found, an alert window opens informing the PLA that unpaid fees were found. - The window will display this text: A people record related to this permit has unpaid ROW fees! - **Stakeholder input needed.** Please provide feedback on the proposed content of the alert window: ![image](https://user-images.githubusercontent.com/49930868/74903386-60935e00-536e-11ea-9440-7aa50dc414c4.png) - The window will facilitate the following actions: - Allow the user to click a button to continue with permit issuance - Allow the user to click a button to abort the permit issuance, and view information related to the qualifying fees - The people invoice list view is an example of how we might provide the user with data about the qualifying fees: ![image](https://user-images.githubusercontent.com/49930868/74903558-e0212d00-536e-11ea-968f-a603023fc067.png) 4. PLA will provide the unpaid fee information (see list view above) to the customer, and note 'Pending Outstanding Fee Payment**' in the folder description. - If it's possible to email the 'list view' of unpaid fees to the customer from the window in AMANDA, that would be ideal **Will need LIT feedback on this.** 5. Customer pays - Jump to # 8 if customer does not pay 6. AMANDA notifies the PLA (or perhaps a resource task list?) that the permit needs to be issued **Stakeholder input needed. How would they most like to be notified?** 7. PLA issues the permit. End Process. 8. **Stakeholder input needed.** Will ROW need/want to know that these fees are still unpaid after a certain amount of time? For example, a monthly report emailed by the system with info about all 'fee flag alerts' that occurred in the previous month; only include info for fees that are still unpaid. - This would be another story/feature however LIT needs to be aware of the reporting requirement when developing the fee flag feature. ------ Old Description: Requesting an alert to notify analysts when there are unpaid fees affiliated with any people records tied to a folder. Something similar to the current alert received when an analyst attempts to issue a permit that has violations affiliated with an address records tied to a folder. Request Date: 2018-11-29 17:12:00 Request ID: DTS18-103080 Status: Backlog Customer Priority: 0 Level of Effort: Small Project DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c00727bd1bf2c4a9f4d268c *Migrated from [atd-amanda #46](https://github.com/cityofaustin/atd-amanda/issues/46)*
1.0
Feature: Unpaid Fee Flag on Permit Issuance - ## Unpaid Fee Flag _AKA # 46 Unpaid Fee Flag on Permit Issuance (Audit Request)_ **Stakeholders:** Paloma, Kim, Mooney **User Story:** As a ROW permit analyst (PLA), I need to know if any of the folder people have unpaid ROW fees BEFORE the permit is issued, so I can place the permit on hold and attempt to collect the outstanding fee(s). **Current State:** 1. The PLA, prior to permit approval, performs manual research to identify any unpaid ROW fees for ALL people listed on the people tab (Contractor, applicant, contact, whoever). - They do not collect fees billed before 10/1/2013. 2. If no outstanding fees are identified, the permit is issued. End process. 3. If outstanding ATD fees are identified, the PLA provides the outstanding fee information to the person/entity responsible for payment. - So if there are 3 unique people/entities on the permit, each entity will be provided with their unique fee information. - The people/entities are provided with instructions for paying, all outstanding bill numbers, and the total for each bill number. 4. The PLA adds a note to the folder description: "Pending Outstanding Fee Payment **TL" 5. The customer will pay the fees, either online or in person. 6. The customer notifies PLA of payment - When this step is missed, it leads to customers working in the ROW without an approved permit 7. PLA issues the permit. End Process. **Future State:** 1. Folder issuance is attempted, either by a user from the folder tab, or by a system procedure like post payment. - We need to be sure any new procedure doesn't cause payment update failures when the fees are paid online. (Put simply, lets learn from the issues caused by the Post Payment upgrade.) - This feature may have to be split into 2 features, one for running this procedure when issuance is initiated by the user from the folder tab, and another for running this procedure when issuance is initiated by a post payment procedure. **Will need LIT feedback on this.** 2. The system performs the unpaid fee verification - See # 1 in Current State; the system needs to do this instead of the PLA - Query all unpaid fees for each people record on the people tab. Of those fees, exclude any not in folder types RW, EX, LC, ECV, GF, or DS. Of those fees, exclude any created prior to 10/1/2013. **Now you have a list of qualifying fees.** 3. If **qualifying fees** are found, an alert window opens informing the PLA that unpaid fees were found. - The window will display this text: A people record related to this permit has unpaid ROW fees! - **Stakeholder input needed.** Please provide feedback on the proposed content of the alert window: ![image](https://user-images.githubusercontent.com/49930868/74903386-60935e00-536e-11ea-9440-7aa50dc414c4.png) - The window will facilitate the following actions: - Allow the user to click a button to continue with permit issuance - Allow the user to click a button to abort the permit issuance, and view information related to the qualifying fees - The people invoice list view is an example of how we might provide the user with data about the qualifying fees: ![image](https://user-images.githubusercontent.com/49930868/74903558-e0212d00-536e-11ea-968f-a603023fc067.png) 4. PLA will provide the unpaid fee information (see list view above) to the customer, and note 'Pending Outstanding Fee Payment**' in the folder description. - If it's possible to email the 'list view' of unpaid fees to the customer from the window in AMANDA, that would be ideal **Will need LIT feedback on this.** 5. Customer pays - Jump to # 8 if customer does not pay 6. AMANDA notifies the PLA (or perhaps a resource task list?) that the permit needs to be issued **Stakeholder input needed. How would they most like to be notified?** 7. PLA issues the permit. End Process. 8. **Stakeholder input needed.** Will ROW need/want to know that these fees are still unpaid after a certain amount of time? For example, a monthly report emailed by the system with info about all 'fee flag alerts' that occurred in the previous month; only include info for fees that are still unpaid. - This would be another story/feature however LIT needs to be aware of the reporting requirement when developing the fee flag feature. ------ Old Description: Requesting an alert to notify analysts when there are unpaid fees affiliated with any people records tied to a folder. Something similar to the current alert received when an analyst attempts to issue a permit that has violations affiliated with an address records tied to a folder. Request Date: 2018-11-29 17:12:00 Request ID: DTS18-103080 Status: Backlog Customer Priority: 0 Level of Effort: Small Project DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c00727bd1bf2c4a9f4d268c *Migrated from [atd-amanda #46](https://github.com/cityofaustin/atd-amanda/issues/46)*
non_process
feature unpaid fee flag on permit issuance unpaid fee flag aka unpaid fee flag on permit issuance audit request stakeholders paloma kim mooney user story as a row permit analyst pla i need to know if any of the folder people have unpaid row fees before the permit is issued so i can place the permit on hold and attempt to collect the outstanding fee s current state the pla prior to permit approval performs manual research to identify any unpaid row fees for all people listed on the people tab contractor applicant contact whoever they do not collect fees billed before if no outstanding fees are identified the permit is issued end process if outstanding atd fees are identified the pla provides the outstanding fee information to the person entity responsible for payment so if there are unique people entities on the permit each entity will be provided with their unique fee information the people entities are provided with instructions for paying all outstanding bill numbers and the total for each bill number the pla adds a note to the folder description pending outstanding fee payment tl the customer will pay the fees either online or in person the customer notifies pla of payment when this step is missed it leads to customers working in the row without an approved permit pla issues the permit end process future state folder issuance is attempted either by a user from the folder tab or by a system procedure like post payment we need to be sure any new procedure doesn t cause payment update failures when the fees are paid online put simply lets learn from the issues caused by the post payment upgrade this feature may have to be split into features one for running this procedure when issuance is initiated by the user from the folder tab and another for running this procedure when issuance is initiated by a post payment procedure will need lit feedback on this the system performs the unpaid fee verification see in current state the system needs to do this instead of the pla query all unpaid fees for each people record on the people tab of those fees exclude any not in folder types rw ex lc ecv gf or ds of those fees exclude any created prior to now you have a list of qualifying fees if qualifying fees are found an alert window opens informing the pla that unpaid fees were found the window will display this text a people record related to this permit has unpaid row fees stakeholder input needed please provide feedback on the proposed content of the alert window the window will facilitate the following actions allow the user to click a button to continue with permit issuance allow the user to click a button to abort the permit issuance and view information related to the qualifying fees the people invoice list view is an example of how we might provide the user with data about the qualifying fees pla will provide the unpaid fee information see list view above to the customer and note pending outstanding fee payment in the folder description if it s possible to email the list view of unpaid fees to the customer from the window in amanda that would be ideal will need lit feedback on this customer pays jump to if customer does not pay amanda notifies the pla or perhaps a resource task list that the permit needs to be issued stakeholder input needed how would they most like to be notified pla issues the permit end process stakeholder input needed will row need want to know that these fees are still unpaid after a certain amount of time for example a monthly report emailed by the system with info about all fee flag alerts that occurred in the previous month only include info for fees that are still unpaid this would be another story feature however lit needs to be aware of the reporting requirement when developing the fee flag feature old description requesting an alert to notify analysts when there are unpaid fees affiliated with any people records tied to a folder something similar to the current alert received when an analyst attempts to issue a permit that has violations affiliated with an address records tied to a folder request date request id status backlog customer priority level of effort small project dts url migrated from
0
21,365
29,194,080,353
IssuesEvent
2023-05-20 00:31:48
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / Caieiras, São Paulo, Brazil] C# Developer (Júnior) (Híbrido) na Coodesh
SALVADOR PJ BANCO DE DADOS FULL-STACK HTML JUNIOR SQL REST SOAP JSON ANGULAR REACT XML REQUISITOS PROCESSOS GITHUB UMA C APIs AUTOMAÇÃO DE PROCESSOS ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong><ins>TRIFORSYS</ins></strong> está em busca de <strong><ins>C# Developer (Júnior)</ins></strong> para integrar seu time!</p> <p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p> <p><strong>Sobre a empresa</strong></p> <p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p> <p></p> <p>Venha com a gente fazer seu negócio!</p> ## Triforsys: <p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a> ## Habilidades: - Angular - HTML - CSS - React.js - Banco de dados relacionais (SQL) ## Local: Caieiras, São Paulo, Brazil ## Requisitos: - Residir em Caieiras ou proximidades (Formato híbrido); - Conhecimentos em APIs, REST, SOAP; - Conhecimentos em XML, JSON, HTML. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C# Developer (Júnior) (Híbrido) na Triforsys](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
2.0
[Hibrido / Caieiras, São Paulo, Brazil] C# Developer (Júnior) (Híbrido) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong><ins>TRIFORSYS</ins></strong> está em busca de <strong><ins>C# Developer (Júnior)</ins></strong> para integrar seu time!</p> <p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p> <p><strong>Sobre a empresa</strong></p> <p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p> <p></p> <p>Venha com a gente fazer seu negócio!</p> ## Triforsys: <p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a> ## Habilidades: - Angular - HTML - CSS - React.js - Banco de dados relacionais (SQL) ## Local: Caieiras, São Paulo, Brazil ## Requisitos: - Residir em Caieiras ou proximidades (Formato híbrido); - Conhecimentos em APIs, REST, SOAP; - Conhecimentos em XML, JSON, HTML. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C# Developer (Júnior) (Híbrido) na Triforsys](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
process
c developer júnior híbrido na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a triforsys está em busca de c developer júnior para integrar seu time fábrica softwares e soluções em ti de forma personalizada atendendo a necessidade do nosso cliente com foco na automação de processos sobre a empresa somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é venha com a gente fazer seu negócio triforsys somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é habilidades angular html css react js banco de dados relacionais sql local caieiras são paulo brazil requisitos residir em caieiras ou proximidades formato híbrido conhecimentos em apis rest soap conhecimentos em xml json html como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria full stack
1
20,426
3,812,781,005
IssuesEvent
2016-03-27 20:52:32
briansmith/ring
https://api.github.com/repos/briansmith/ring
closed
Remove all code that supports AES-192
enhancement good-first-bug performance static-analysis-and-type-safety test-coverage
See https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551 and https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551 Also, I guess the assembly language code must have some bits that support 192-bit AES. We don't need any of this since we don't (and don't plan to) expose AES-192.
1.0
Remove all code that supports AES-192 - See https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551 and https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551 Also, I guess the assembly language code must have some bits that support 192-bit AES. We don't need any of this since we don't (and don't plan to) expose AES-192.
non_process
remove all code that supports aes see and also i guess the assembly language code must have some bits that support bit aes we don t need any of this since we don t and don t plan to expose aes
0
6,147
2,814,174,529
IssuesEvent
2015-05-18 18:32:45
joyent/node
https://api.github.com/repos/joyent/node
closed
test: gc/test-net-timeout fails consistently
test
``` Done: 500/500 Collected: 499/500 All should be collected now. Collected: 499/500 timers.js:102 if (!process.listeners('uncaughtException').length) throw e; ^ AssertionError: false == true at null._onTimeout (/home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js:64:7) at Timer.list.ontimeout (timers.js:100:19) at process._makeCallback (node.js:248:20) Command: out/Release/node --expose-gc /home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js ``` /cc @isaacs?
1.0
test: gc/test-net-timeout fails consistently - ``` Done: 500/500 Collected: 499/500 All should be collected now. Collected: 499/500 timers.js:102 if (!process.listeners('uncaughtException').length) throw e; ^ AssertionError: false == true at null._onTimeout (/home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js:64:7) at Timer.list.ontimeout (timers.js:100:19) at process._makeCallback (node.js:248:20) Command: out/Release/node --expose-gc /home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js ``` /cc @isaacs?
non_process
test gc test net timeout fails consistently done collected all should be collected now collected timers js if process listeners uncaughtexception length throw e assertionerror false true at null ontimeout home bnoordhuis src nodejs master test gc test net timeout js at timer list ontimeout timers js at process makecallback node js command out release node expose gc home bnoordhuis src nodejs master test gc test net timeout js cc isaacs
0
371,868
10,982,278,821
IssuesEvent
2019-12-01 06:03:33
Luna-Interactive/catastrophe
https://api.github.com/repos/Luna-Interactive/catastrophe
closed
Winning
Priority: Critical SFX
**Is your feature request related to a problem? Please describe.** Missing sound for when a player wins the game.
1.0
Winning - **Is your feature request related to a problem? Please describe.** Missing sound for when a player wins the game.
non_process
winning is your feature request related to a problem please describe missing sound for when a player wins the game
0
5,241
8,036,963,753
IssuesEvent
2018-07-30 10:55:24
Open-EO/openeo-api
https://api.github.com/repos/Open-EO/openeo-api
closed
Variables in process_graphs?
process graph management processes vote work in progress
In issue #52 about product name differences, I had the idea that stored process graphs, i.e. /process_graphs/:id, may need some variables in them to be really portable. Otherwise sharing as proposed in #85 might not be so useful. An example would be the product id in case we can't come up with something useful to solve the product name differences. Process graphs wouldn't be portable in case a back-end specific product name would be stored with the process_graph. Changes to `POST /process_graphs` (and similarly `PATCH /process_graphs/:id`) could look like this: ``` { "process_graph":{ "process_id":"min_time", "description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.", "imagery":{ "process_id":"NDVI", "imagery":{ "process_id":"filter_daterange", "imagery":{ "process_id":"get_data", "data_id":"{$product}" }, "from":"2017-01-01", "to":"2017-01-31" }, "red":"{$red_band}", "nir":"{$nir_band}" } }, "variables":{ "product":{ "type":"string", "default":"Sentinel-2A", "description":"Product id of the dataset" }, "red_band":{ "type":"string", "default":"B4", "description":"Red Band id of the dataset" }, "nir_band":{ "type":"string", "default":"B8", "description":"NIR Band id of the dataset" } } } ``` `"{$product}"`, `"{$red_band}"` and `"{$nir_band}"` are the variables to be replaced - including the quotes to keep it being valid, parsable JSON. We could an additional (but optional) property "variables" that could hold the data type, default values and a description. Types could be: everything that is a valid JSON data type, including objects and arrays. Alternatively, we could make it look like JSON schema: ``` { "process_graph":{ "process_id":"min_time", "description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.", "imagery":{ "process_id":"NDVI", "imagery":{ "process_id":"filter_daterange", "imagery":{ "process_id":"get_data", "data_id":{ "$ref":"product", "type":"string", "default":"Sentinel-2A", "description":"Product id of the dataset" } }, "from":"2017-01-01", "to":"2017-01-31" }, "red":{ "$ref":"red_band", "type":"string", "default":"B4", "description":"Red Band id of the dataset" }, "nir":{ "$ref":"nir_band", "type":"string", "default":"B8", "description":"NIR Band id of the dataset" } } } } ``` That approach allows for better encapsulation. The "$ref" key would be the "identifier" for the variables JSON object. "$ref" could also be "variable_id" to be more in-line with "process_id". `GET /process_graphs/:id` could work in two modes: 1. Variables could be transferred **serialized** as JSON via query parameters and the back-end returns the resulting process graph 2. The requesting entities processes/replaces the variables. Example for the first approach: Example request: `GET https://www.openeo.org/api/v0/process_graphs/1234?product=%22Sentinel-2A-L1C%22` The product would be changed and the bands would be taken from the specified defaults. Resulting process_graph: ``` { "process_id":"min_time", "description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.", "imagery":{ "process_id":"NDVI", "imagery":{ "process_id":"filter_daterange", "imagery":{ "process_id":"get_data", "data_id":"Sentinal-2A-L1C" }, "from":"2017-01-01", "to":"2017-01-31" }, "red":"B4", "nir":"B8" } } ``` Both modes could be used in a process like this: ``` { "process_id":"process_graph", "url":"https://www.openeo.org/api/v0/process_graphs/1234", "product":"Sentinal-2A-L1C" } ```
2.0
Variables in process_graphs? - In issue #52 about product name differences, I had the idea that stored process graphs, i.e. /process_graphs/:id, may need some variables in them to be really portable. Otherwise sharing as proposed in #85 might not be so useful. An example would be the product id in case we can't come up with something useful to solve the product name differences. Process graphs wouldn't be portable in case a back-end specific product name would be stored with the process_graph. Changes to `POST /process_graphs` (and similarly `PATCH /process_graphs/:id`) could look like this: ``` { "process_graph":{ "process_id":"min_time", "description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.", "imagery":{ "process_id":"NDVI", "imagery":{ "process_id":"filter_daterange", "imagery":{ "process_id":"get_data", "data_id":"{$product}" }, "from":"2017-01-01", "to":"2017-01-31" }, "red":"{$red_band}", "nir":"{$nir_band}" } }, "variables":{ "product":{ "type":"string", "default":"Sentinel-2A", "description":"Product id of the dataset" }, "red_band":{ "type":"string", "default":"B4", "description":"Red Band id of the dataset" }, "nir_band":{ "type":"string", "default":"B8", "description":"NIR Band id of the dataset" } } } ``` `"{$product}"`, `"{$red_band}"` and `"{$nir_band}"` are the variables to be replaced - including the quotes to keep it being valid, parsable JSON. We could an additional (but optional) property "variables" that could hold the data type, default values and a description. Types could be: everything that is a valid JSON data type, including objects and arrays. Alternatively, we could make it look like JSON schema: ``` { "process_graph":{ "process_id":"min_time", "description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.", "imagery":{ "process_id":"NDVI", "imagery":{ "process_id":"filter_daterange", "imagery":{ "process_id":"get_data", "data_id":{ "$ref":"product", "type":"string", "default":"Sentinel-2A", "description":"Product id of the dataset" } }, "from":"2017-01-01", "to":"2017-01-31" }, "red":{ "$ref":"red_band", "type":"string", "default":"B4", "description":"Red Band id of the dataset" }, "nir":{ "$ref":"nir_band", "type":"string", "default":"B8", "description":"NIR Band id of the dataset" } } } } ``` That approach allows for better encapsulation. The "$ref" key would be the "identifier" for the variables JSON object. "$ref" could also be "variable_id" to be more in-line with "process_id". `GET /process_graphs/:id` could work in two modes: 1. Variables could be transferred **serialized** as JSON via query parameters and the back-end returns the resulting process graph 2. The requesting entities processes/replaces the variables. Example for the first approach: Example request: `GET https://www.openeo.org/api/v0/process_graphs/1234?product=%22Sentinel-2A-L1C%22` The product would be changed and the bands would be taken from the specified defaults. Resulting process_graph: ``` { "process_id":"min_time", "description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.", "imagery":{ "process_id":"NDVI", "imagery":{ "process_id":"filter_daterange", "imagery":{ "process_id":"get_data", "data_id":"Sentinal-2A-L1C" }, "from":"2017-01-01", "to":"2017-01-31" }, "red":"B4", "nir":"B8" } } ``` Both modes could be used in a process like this: ``` { "process_id":"process_graph", "url":"https://www.openeo.org/api/v0/process_graphs/1234", "product":"Sentinal-2A-L1C" } ```
process
variables in process graphs in issue about product name differences i had the idea that stored process graphs i e process graphs id may need some variables in them to be really portable otherwise sharing as proposed in might not be so useful an example would be the product id in case we can t come up with something useful to solve the product name differences process graphs wouldn t be portable in case a back end specific product name would be stored with the process graph changes to post process graphs and similarly patch process graphs id could look like this process graph process id min time description deriving minimum ndvi measurements over pixel time series of sentinel imagery imagery process id ndvi imagery process id filter daterange imagery process id get data data id product from to red red band nir nir band variables product type string default sentinel description product id of the dataset red band type string default description red band id of the dataset nir band type string default description nir band id of the dataset product red band and nir band are the variables to be replaced including the quotes to keep it being valid parsable json we could an additional but optional property variables that could hold the data type default values and a description types could be everything that is a valid json data type including objects and arrays alternatively we could make it look like json schema process graph process id min time description deriving minimum ndvi measurements over pixel time series of sentinel imagery imagery process id ndvi imagery process id filter daterange imagery process id get data data id ref product type string default sentinel description product id of the dataset from to red ref red band type string default description red band id of the dataset nir ref nir band type string default description nir band id of the dataset that approach allows for better encapsulation the ref key would be the identifier for the variables json object ref could also be variable id to be more in line with process id get process graphs id could work in two modes variables could be transferred serialized as json via query parameters and the back end returns the resulting process graph the requesting entities processes replaces the variables example for the first approach example request get the product would be changed and the bands would be taken from the specified defaults resulting process graph process id min time description deriving minimum ndvi measurements over pixel time series of sentinel imagery imagery process id ndvi imagery process id filter daterange imagery process id get data data id sentinal from to red nir both modes could be used in a process like this process id process graph url product sentinal
1
10,118
13,044,162,241
IssuesEvent
2020-07-29 03:47:31
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `GetFormat` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `GetFormat` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `GetFormat` from TiDB - ## Description Port the scalar function `GetFormat` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function getformat from tidb description port the scalar function getformat from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
1
13,102
15,496,467,659
IssuesEvent
2021-03-11 02:44:05
dluiscosta/weather_api
https://api.github.com/repos/dluiscosta/weather_api
opened
Establish test dependency hierarchy
development process enhancement
Establish test dependency hierarchy where some test ```A``` might assert the proper operation of a feature, which in turn requires the proper operation of a second feature, asserted independently by some test ```B```, thus making ```A``` dependent on ```B```. That way, when ```B``` fails, ```A``` can be skipped (as it's bound to fail as well), thus saving time.
1.0
Establish test dependency hierarchy - Establish test dependency hierarchy where some test ```A``` might assert the proper operation of a feature, which in turn requires the proper operation of a second feature, asserted independently by some test ```B```, thus making ```A``` dependent on ```B```. That way, when ```B``` fails, ```A``` can be skipped (as it's bound to fail as well), thus saving time.
process
establish test dependency hierarchy establish test dependency hierarchy where some test a might assert the proper operation of a feature which in turn requires the proper operation of a second feature asserted independently by some test b thus making a dependent on b that way when b fails a can be skipped as it s bound to fail as well thus saving time
1
18,067
24,080,244,816
IssuesEvent
2022-09-19 05:32:02
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing algorithm "Lines to polygons" produces incorrect polygons
Processing Bug
### What is the bug or the crash? When using the processing algorithm "Lines to polygons" with certain data, it produces incorrect polygons. Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip) ### Steps to reproduce the issue 1. Add lines_simplified.json to the project ![image](https://user-images.githubusercontent.com/33314057/190582982-f534d66e-2e9f-451d-a2e2-e2f2ba0608c4.png) 2. Run "Lines to polygons" processing algorithm with the layer ![image](https://user-images.githubusercontent.com/33314057/190583033-87ac8d31-06ae-4c97-9747-f17dcd9778bf.png) ### Versions QGISin versio 3.22.9-Białowieża QGIS-koodin versio a8e9e6fa Qt:n versio 5.15.3 Python-versio 3.9.5 GDAL/OGR:n versio 3.5.1 PROJ-versio 9.0.1 EPSG-rekisteritietokannan versio v10.064 (2022-05-19) GEOS-versio 3.10.3-CAPI-1.16.1 SQLite-versio 3.38.1 PDAL-versio 2.3.0 PostgreSQL:n asiakkaan versio 14.3 SpatiaLite-versio 5.0.1 QWT-versio 6.1.6 QScintilla2-versio 2.13.1 Käyttöjärjestelmän versio Windows 10 Version 2009 Aktiiviset python-lisäosat db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 plugin_reloader 0.9.1 Ilre 0.1.0 Also tested with QGIS 3.26.2 on Arch Linux ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context The problem might lie with the dataset used. The same dataset causes the problems in #50227
1.0
Processing algorithm "Lines to polygons" produces incorrect polygons - ### What is the bug or the crash? When using the processing algorithm "Lines to polygons" with certain data, it produces incorrect polygons. Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip) ### Steps to reproduce the issue 1. Add lines_simplified.json to the project ![image](https://user-images.githubusercontent.com/33314057/190582982-f534d66e-2e9f-451d-a2e2-e2f2ba0608c4.png) 2. Run "Lines to polygons" processing algorithm with the layer ![image](https://user-images.githubusercontent.com/33314057/190583033-87ac8d31-06ae-4c97-9747-f17dcd9778bf.png) ### Versions QGISin versio 3.22.9-Białowieża QGIS-koodin versio a8e9e6fa Qt:n versio 5.15.3 Python-versio 3.9.5 GDAL/OGR:n versio 3.5.1 PROJ-versio 9.0.1 EPSG-rekisteritietokannan versio v10.064 (2022-05-19) GEOS-versio 3.10.3-CAPI-1.16.1 SQLite-versio 3.38.1 PDAL-versio 2.3.0 PostgreSQL:n asiakkaan versio 14.3 SpatiaLite-versio 5.0.1 QWT-versio 6.1.6 QScintilla2-versio 2.13.1 Käyttöjärjestelmän versio Windows 10 Version 2009 Aktiiviset python-lisäosat db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 plugin_reloader 0.9.1 Ilre 0.1.0 Also tested with QGIS 3.26.2 on Arch Linux ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context The problem might lie with the dataset used. The same dataset causes the problems in #50227
process
processing algorithm lines to polygons produces incorrect polygons what is the bug or the crash when using the processing algorithm lines to polygons with certain data it produces incorrect polygons here is a simple topologically correct line layer to test the bug with steps to reproduce the issue add lines simplified json to the project run lines to polygons processing algorithm with the layer versions qgisin versio białowieża qgis koodin versio qt n versio python versio gdal ogr n versio proj versio epsg rekisteritietokannan versio geos versio capi sqlite versio pdal versio postgresql n asiakkaan versio spatialite versio qwt versio versio käyttöjärjestelmän versio windows version aktiiviset python lisäosat db manager grassprovider metasearch processing sagaprovider plugin reloader ilre also tested with qgis on arch linux supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context the problem might lie with the dataset used the same dataset causes the problems in
1
408,546
27,695,438,910
IssuesEvent
2023-03-14 01:33:12
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
[Gitbook Documentation] Complete 'Extracting Location Coordinates to add a Map Marker Map' article
Workgroup: DTS Service: Apps Type: Documentation Product: TDS Portal
As exemplified in TDS or S&M
1.0
[Gitbook Documentation] Complete 'Extracting Location Coordinates to add a Map Marker Map' article - As exemplified in TDS or S&M
non_process
complete extracting location coordinates to add a map marker map article as exemplified in tds or s m
0
19,615
25,970,594,587
IssuesEvent
2022-12-19 10:54:07
toggl/track-windows-feedback
https://api.github.com/repos/toggl/track-windows-feedback
closed
Is there a Beta "Release channel"?
solved processed
I see the option for changing to the Beta "Release channel" but when I select it and then close the window and re-open it switches back to the Stable channel. Is there a Beta channel anymore? I know in the past (old app) there was and I was on the Beta channel but not sure if now with new native app if maybe there is only a Stable channel and this dropdown should be ignored? Thanks! ![image](https://user-images.githubusercontent.com/64155612/206004480-fcfb6b6a-f91e-4014-a096-c02291b02ce1.png)
1.0
Is there a Beta "Release channel"? - I see the option for changing to the Beta "Release channel" but when I select it and then close the window and re-open it switches back to the Stable channel. Is there a Beta channel anymore? I know in the past (old app) there was and I was on the Beta channel but not sure if now with new native app if maybe there is only a Stable channel and this dropdown should be ignored? Thanks! ![image](https://user-images.githubusercontent.com/64155612/206004480-fcfb6b6a-f91e-4014-a096-c02291b02ce1.png)
process
is there a beta release channel i see the option for changing to the beta release channel but when i select it and then close the window and re open it switches back to the stable channel is there a beta channel anymore i know in the past old app there was and i was on the beta channel but not sure if now with new native app if maybe there is only a stable channel and this dropdown should be ignored thanks
1
37,342
15,262,770,210
IssuesEvent
2021-02-22 00:38:34
Azure/azure-sdk-for-net
https://api.github.com/repos/Azure/azure-sdk-for-net
closed
Error while using Azure.Messaging.ServiceBus library for Service bus trigger function
Client Functions Service Bus customer-reported needs-author-feedback question
Azure.Messaging.ServiceBus 7.1.0 Microsoft.Azure.WebJobs.Extensions.ServiceBus 4.2.1 Microsoft.NET.Sdk.Functions 3.0.11 Azure Cloud. Windows/Linux VS 2019 Hello I have an Azure function, with a Service Bus trigger, using Azure.Messaging.ServiceBus; public static void Run(ServiceBusReceivedMessage myQueueItem, ILogger log, Microsoft.Azure.WebJobs.ExecutionContext context) {..} Is that supported? The [documentation](https://docs.microsoft.com/en-us/azure/azure-functions/functions-bindings-service-bus-trigger?tabs=csharp#usage) says that it should be used Microsoft.Azure.ServiceBus libray, but I need to used the newest libray. The code throws an error so it doesn't seem to be supported.. Binding parameters to complex objects (such as 'ServiceBusReceivedMessage') uses Json.NET serialization or XML object serialization. 1. If ContentType is 'application/json' deserialize as JSON 2. If ContentType is not 'application/json' attempt to deserialize using Message.GetBody, which will handle cases like XML object serialization 3. If this deserialization fails, do a final attempt at JSON deserialization to catch cases where the content type might be incorrect The JSON parser failed: Unexpected character encountered while parsing value: T. Path '', line 0, position 0. Thanks.
1.0
Error while using Azure.Messaging.ServiceBus library for Service bus trigger function - Azure.Messaging.ServiceBus 7.1.0 Microsoft.Azure.WebJobs.Extensions.ServiceBus 4.2.1 Microsoft.NET.Sdk.Functions 3.0.11 Azure Cloud. Windows/Linux VS 2019 Hello I have an Azure function, with a Service Bus trigger, using Azure.Messaging.ServiceBus; public static void Run(ServiceBusReceivedMessage myQueueItem, ILogger log, Microsoft.Azure.WebJobs.ExecutionContext context) {..} Is that supported? The [documentation](https://docs.microsoft.com/en-us/azure/azure-functions/functions-bindings-service-bus-trigger?tabs=csharp#usage) says that it should be used Microsoft.Azure.ServiceBus libray, but I need to used the newest libray. The code throws an error so it doesn't seem to be supported.. Binding parameters to complex objects (such as 'ServiceBusReceivedMessage') uses Json.NET serialization or XML object serialization. 1. If ContentType is 'application/json' deserialize as JSON 2. If ContentType is not 'application/json' attempt to deserialize using Message.GetBody, which will handle cases like XML object serialization 3. If this deserialization fails, do a final attempt at JSON deserialization to catch cases where the content type might be incorrect The JSON parser failed: Unexpected character encountered while parsing value: T. Path '', line 0, position 0. Thanks.
non_process
error while using azure messaging servicebus library for service bus trigger function azure messaging servicebus microsoft azure webjobs extensions servicebus microsoft net sdk functions azure cloud windows linux vs hello i have an azure function with a service bus trigger using azure messaging servicebus public static void run servicebusreceivedmessage myqueueitem ilogger log microsoft azure webjobs executioncontext context is that supported the says that it should be used microsoft azure servicebus libray but i need to used the newest libray the code throws an error so it doesn t seem to be supported binding parameters to complex objects such as servicebusreceivedmessage uses json net serialization or xml object serialization if contenttype is application json deserialize as json if contenttype is not application json attempt to deserialize using message getbody which will handle cases like xml object serialization if this deserialization fails do a final attempt at json deserialization to catch cases where the content type might be incorrect the json parser failed unexpected character encountered while parsing value t path line position thanks
0
217,525
16,855,802,384
IssuesEvent
2021-06-21 06:27:36
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
raftstore::test_merge::test_node_merge_prerequisites_check failed
component/test-bench
raftstore::test_merge::test_node_merge_prerequisites_check Latest failed builds: https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/18842/display/redirect
1.0
raftstore::test_merge::test_node_merge_prerequisites_check failed - raftstore::test_merge::test_node_merge_prerequisites_check Latest failed builds: https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/18842/display/redirect
non_process
raftstore test merge test node merge prerequisites check failed raftstore test merge test node merge prerequisites check latest failed builds
0
719,588
24,764,745,723
IssuesEvent
2022-10-22 11:17:00
DISSINET/InkVisitor
https://api.github.com/repos/DISSINET/InkVisitor
closed
New reference; changing the parsing notation for references
data parsing priority
1. Please include a new col. in Actions and Concepts into the parsing: wordnet_sense_key. 2. I had an inconsistency in the notation of wordnet cols between Actions and Concepts ("special" vs. "reference"). Now unified to "reference". 3. In cols. to be parsed as "reference", I now added the Resource ID (R entity) to the row no. 3, and created those entities in the normal [list of Rs](https://docs.google.com/spreadsheets/d/13eVorFf7J9R8YzO7TmJRVLzIIwRJS737r7eFbH1boyE/edit#gid=583750775).
1.0
New reference; changing the parsing notation for references - 1. Please include a new col. in Actions and Concepts into the parsing: wordnet_sense_key. 2. I had an inconsistency in the notation of wordnet cols between Actions and Concepts ("special" vs. "reference"). Now unified to "reference". 3. In cols. to be parsed as "reference", I now added the Resource ID (R entity) to the row no. 3, and created those entities in the normal [list of Rs](https://docs.google.com/spreadsheets/d/13eVorFf7J9R8YzO7TmJRVLzIIwRJS737r7eFbH1boyE/edit#gid=583750775).
non_process
new reference changing the parsing notation for references please include a new col in actions and concepts into the parsing wordnet sense key i had an inconsistency in the notation of wordnet cols between actions and concepts special vs reference now unified to reference in cols to be parsed as reference i now added the resource id r entity to the row no and created those entities in the normal
0
113,790
4,569,032,192
IssuesEvent
2016-09-15 16:02:01
isobar-techchallenge/isobar-a9f8d858-fb06-4865-833f-f03fc2d03cd1
https://api.github.com/repos/isobar-techchallenge/isobar-a9f8d858-fb06-4865-833f-f03fc2d03cd1
opened
Target Button in Email to Small
priority:low
Within email to user the CTA button target space is to small. Need to make the target click area the full width of the button.
1.0
Target Button in Email to Small - Within email to user the CTA button target space is to small. Need to make the target click area the full width of the button.
non_process
target button in email to small within email to user the cta button target space is to small need to make the target click area the full width of the button
0
18,148
24,187,126,890
IssuesEvent
2022-09-23 14:12:03
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
closed
Golang Integration Tests Failing
bug development-process
**Describe the bug** <!-- A clear and concise description of what the bug is. --> The arrow integration tests are failing on master, https://github.com/apache/arrow-rs/runs/8249872730?check_suite_focus=true, this appears to be caused by https://github.com/apache/arrow/pull/14067 which bumps the minimum required Golang version to 1.17. Unfortunately the version in `apache/arrow-dev:amd64-conda-integration` is 1.16 **To Reproduce** <!-- Steps to reproduce the behavior: --> **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> **Additional context** <!-- Add any other context about the problem here. -->
1.0
Golang Integration Tests Failing - **Describe the bug** <!-- A clear and concise description of what the bug is. --> The arrow integration tests are failing on master, https://github.com/apache/arrow-rs/runs/8249872730?check_suite_focus=true, this appears to be caused by https://github.com/apache/arrow/pull/14067 which bumps the minimum required Golang version to 1.17. Unfortunately the version in `apache/arrow-dev:amd64-conda-integration` is 1.16 **To Reproduce** <!-- Steps to reproduce the behavior: --> **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> **Additional context** <!-- Add any other context about the problem here. -->
process
golang integration tests failing describe the bug a clear and concise description of what the bug is the arrow integration tests are failing on master this appears to be caused by which bumps the minimum required golang version to unfortunately the version in apache arrow dev conda integration is to reproduce steps to reproduce the behavior expected behavior a clear and concise description of what you expected to happen additional context add any other context about the problem here
1
445,518
12,832,125,680
IssuesEvent
2020-07-07 07:04:31
Automattic/abacus
https://api.github.com/repos/Automattic/abacus
opened
Apply consistent object normalisation across the codebase
[!priority] medium [component] experimenter interface [type] enhancement
Mapping nested objects (e.g., metrics to an experiment's metric assignments) is currently handled inconsistently throughout the code. @jessie-ross suggested we use [normalizr](https://github.com/paularmstrong/normalizr) for this. See: https://github.com/Automattic/abacus/pull/198#discussion_r449975708
1.0
Apply consistent object normalisation across the codebase - Mapping nested objects (e.g., metrics to an experiment's metric assignments) is currently handled inconsistently throughout the code. @jessie-ross suggested we use [normalizr](https://github.com/paularmstrong/normalizr) for this. See: https://github.com/Automattic/abacus/pull/198#discussion_r449975708
non_process
apply consistent object normalisation across the codebase mapping nested objects e g metrics to an experiment s metric assignments is currently handled inconsistently throughout the code jessie ross suggested we use for this see
0
3,666
6,694,824,296
IssuesEvent
2017-10-10 04:46:39
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
opened
View Watering Assignments - Finished Assignment Upload
enhancement process workflow
Create a data uploader to upload the finished assignment upload.
1.0
View Watering Assignments - Finished Assignment Upload - Create a data uploader to upload the finished assignment upload.
process
view watering assignments finished assignment upload create a data uploader to upload the finished assignment upload
1
57,769
14,219,807,762
IssuesEvent
2020-11-17 13:47:01
Automattic/wp-calypso
https://api.github.com/repos/Automattic/wp-calypso
closed
Modularise `userDevices`
Build
The `userDevices` portion of the state tree needs to be modularised. See the [modularised state documentation](https://github.com/Automattic/wp-calypso/blob/master/docs/modularized-state.md) for more details.
1.0
Modularise `userDevices` - The `userDevices` portion of the state tree needs to be modularised. See the [modularised state documentation](https://github.com/Automattic/wp-calypso/blob/master/docs/modularized-state.md) for more details.
non_process
modularise userdevices the userdevices portion of the state tree needs to be modularised see the for more details
0
13,338
15,800,886,211
IssuesEvent
2021-04-03 01:47:32
PyCQA/flake8
https://api.github.com/repos/PyCQA/flake8
closed
Order is nondeterministic with multiprocessing and piped output
bug:confirmed component:docs component:multiprocessing feature:accepted fix:committed fix:released
In GitLab by @quentinp on Nov 17, 2014, 03:24 Steps to reproduce: * In a directory, create two files a.py and b.py with an error each (eg. both containing a=1). * From the directory, run `flake8 -j2 . | tee` multiple times (``for i in `seq 1 10` ; do flake8 -j2 . | tee; echo; done`` can be useful). Expected result: The order is always the same, eg. `a.py`, then `b.py`. This is the case with `pep8` and `flake8 -j1`, or when the output is not piped. (By the way, I originally hit the error with `watch`, not `tee`). Actual result: The order is nondeterministic: sometimes `a.py` is first, sometimes `b.py` is first instead.
1.0
Order is nondeterministic with multiprocessing and piped output - In GitLab by @quentinp on Nov 17, 2014, 03:24 Steps to reproduce: * In a directory, create two files a.py and b.py with an error each (eg. both containing a=1). * From the directory, run `flake8 -j2 . | tee` multiple times (``for i in `seq 1 10` ; do flake8 -j2 . | tee; echo; done`` can be useful). Expected result: The order is always the same, eg. `a.py`, then `b.py`. This is the case with `pep8` and `flake8 -j1`, or when the output is not piped. (By the way, I originally hit the error with `watch`, not `tee`). Actual result: The order is nondeterministic: sometimes `a.py` is first, sometimes `b.py` is first instead.
process
order is nondeterministic with multiprocessing and piped output in gitlab by quentinp on nov steps to reproduce in a directory create two files a py and b py with an error each eg both containing a from the directory run tee multiple times for i in seq do tee echo done can be useful expected result the order is always the same eg a py then b py this is the case with and or when the output is not piped by the way i originally hit the error with watch not tee actual result the order is nondeterministic sometimes a py is first sometimes b py is first instead
1
42,271
9,199,573,387
IssuesEvent
2019-03-07 15:13:47
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Buttons contributed to the extensions toolbar should allow to specify "toggled" state
code-nav extensions feature-request
Many buttons contributed to the toolbar are to toggle something on or off, e.g. Coverage colors, soon code intel, ... The on/off state can currently only be communicated through the label "Enable"/"Disable". This makes it impossible to omit the lable to save space, and looks very weird on GitHub where the buttons look like actual 3D buttons, that don't change into a "pressed" state when the setting is enabled. There should be a field `"toggled"` in package.json for the button that is a boolean expression (and can read a setting for example) that makes the button toggled or not. On the Sourcegraph webapp, the state could be indicated by a background color change (something like `rgba(0, 0, 0, 0.5)` in light theme, `rgba(255, 255, 255, 0.5)` in dark theme). I see this as a prerequisite to add an enable/disable for TypeScript because otherwise the experience is very weird. Rel #1572
1.0
Buttons contributed to the extensions toolbar should allow to specify "toggled" state - Many buttons contributed to the toolbar are to toggle something on or off, e.g. Coverage colors, soon code intel, ... The on/off state can currently only be communicated through the label "Enable"/"Disable". This makes it impossible to omit the lable to save space, and looks very weird on GitHub where the buttons look like actual 3D buttons, that don't change into a "pressed" state when the setting is enabled. There should be a field `"toggled"` in package.json for the button that is a boolean expression (and can read a setting for example) that makes the button toggled or not. On the Sourcegraph webapp, the state could be indicated by a background color change (something like `rgba(0, 0, 0, 0.5)` in light theme, `rgba(255, 255, 255, 0.5)` in dark theme). I see this as a prerequisite to add an enable/disable for TypeScript because otherwise the experience is very weird. Rel #1572
non_process
buttons contributed to the extensions toolbar should allow to specify toggled state many buttons contributed to the toolbar are to toggle something on or off e g coverage colors soon code intel the on off state can currently only be communicated through the label enable disable this makes it impossible to omit the lable to save space and looks very weird on github where the buttons look like actual buttons that don t change into a pressed state when the setting is enabled there should be a field toggled in package json for the button that is a boolean expression and can read a setting for example that makes the button toggled or not on the sourcegraph webapp the state could be indicated by a background color change something like rgba in light theme rgba in dark theme i see this as a prerequisite to add an enable disable for typescript because otherwise the experience is very weird rel
0
13,241
15,708,282,161
IssuesEvent
2021-03-26 20:16:17
xatkit-bot-platform/xatkit-runtime
https://api.github.com/repos/xatkit-bot-platform/xatkit-runtime
opened
Remove @Ignore for language detection tests
Processors Testing
We have to ignore these tests for the moment because we don't have the infrastructure to run them in a CI/CD environment.
1.0
Remove @Ignore for language detection tests - We have to ignore these tests for the moment because we don't have the infrastructure to run them in a CI/CD environment.
process
remove ignore for language detection tests we have to ignore these tests for the moment because we don t have the infrastructure to run them in a ci cd environment
1
5,108
7,885,450,349
IssuesEvent
2018-06-27 12:31:08
Open-EO/openeo-api
https://api.github.com/repos/Open-EO/openeo-api
opened
Debugging and getting intermediate result and metadata
in discussion other process graphs processes
A big topic during the hackathon was debugging of our process graphs and getting information of intermediate result and other meta data. For example, after a filter one might want to know what data is left. GEE allows to add print() calls to get information about a variable with results between different steps. How could something like that be made available in openEO?
2.0
Debugging and getting intermediate result and metadata - A big topic during the hackathon was debugging of our process graphs and getting information of intermediate result and other meta data. For example, after a filter one might want to know what data is left. GEE allows to add print() calls to get information about a variable with results between different steps. How could something like that be made available in openEO?
process
debugging and getting intermediate result and metadata a big topic during the hackathon was debugging of our process graphs and getting information of intermediate result and other meta data for example after a filter one might want to know what data is left gee allows to add print calls to get information about a variable with results between different steps how could something like that be made available in openeo
1
288,556
24,915,737,836
IssuesEvent
2022-10-30 11:29:29
Jerrylum/ProtocolDiagram
https://api.github.com/repos/Jerrylum/ProtocolDiagram
closed
Generate diagram unit cases
test
I read the code on branch handle-diagram-generation. I see it is canvas-based. How are you going to unit test the code? Because I think comparing the canvas content in 2D in the test cases is not a good practice. @JohnDoeAntler
1.0
Generate diagram unit cases - I read the code on branch handle-diagram-generation. I see it is canvas-based. How are you going to unit test the code? Because I think comparing the canvas content in 2D in the test cases is not a good practice. @JohnDoeAntler
non_process
generate diagram unit cases i read the code on branch handle diagram generation i see it is canvas based how are you going to unit test the code because i think comparing the canvas content in in the test cases is not a good practice johndoeantler
0
500,247
14,494,248,168
IssuesEvent
2020-12-11 09:35:07
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
rule34.xxx - design is broken
browser-fenix engine-gecko priority-normal
<!-- @browser: Firefox Mobile 85.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/63472 --> <!-- @extra_labels: browser-fenix --> **URL**: https://rule34.xxx/index.php?page=post&s=list&tags=sinensian+ **Browser / Version**: Firefox Mobile 85.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Images not loaded **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206093054</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/8fc9e360-40b1-4a50-9f35-e288f86d8736) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
rule34.xxx - design is broken - <!-- @browser: Firefox Mobile 85.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/63472 --> <!-- @extra_labels: browser-fenix --> **URL**: https://rule34.xxx/index.php?page=post&s=list&tags=sinensian+ **Browser / Version**: Firefox Mobile 85.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Images not loaded **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206093054</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/8fc9e360-40b1-4a50-9f35-e288f86d8736) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
xxx design is broken url browser version firefox mobile operating system android tested another browser yes chrome problem type design is broken description images not loaded steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
93,754
3,910,681,267
IssuesEvent
2016-04-20 00:13:41
sdsu-its/fit-welcome
https://api.github.com/repos/sdsu-its/fit-welcome
closed
Cannot Swipe Card
Priority-Minor Type-Bug
When trying to login a user via a MagStripe Reader, the browser rejects the entry because it does not follow the specified format.
1.0
Cannot Swipe Card - When trying to login a user via a MagStripe Reader, the browser rejects the entry because it does not follow the specified format.
non_process
cannot swipe card when trying to login a user via a magstripe reader the browser rejects the entry because it does not follow the specified format
0
81,419
15,726,672,837
IssuesEvent
2021-03-29 11:39:04
GTNewHorizons/GT-New-Horizons-Modpack
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
closed
Evaluate multiblock ore drill speed/rebalance fluid use
Status: RFC (request for comment) Status: stale Type: Need Code changes
Multiblock ore drill speed is not very fast. Players would like to speed it up. Should it compare to Ender Quarry? Also, would be interesting to have more fluid options. So, suggestion: ~~- [ ] Reduce fluid use at the drill by 10x. Increase the lube/stone dust use in the mixer by 10x to compensate. This way players can use 256B drums to provide drilling fluid instead of needing bedrockium drums.~~ Players seem ok with this as-is. - [ ] Increase drill base speed. For tier 2, base speed is 2s, adjust up/down appropriately for other tiers. Base speed - need to determine how this applies per tier. Should it all be respect to LV? Or LV = tier 1, MV = tier 2, etc? - [ ] Add API to add custom advanced fluids, and how much faster they are compared to default. IE, fluid.advlube or whatever. Could be 4x faster, maybe add higher tier options. - [ ] Let overclocking dictate base speed adjustments. Let advanced fluid speeds dictate how many blocks are picked per operation. Basic drilling fluid = 1x. Advanced = 4x. Maybe faster ones? - [ ] Add a second fluid hatch for concrete? Let the machine place concrete in the drill holes. Dunno if that matters.
1.0
Evaluate multiblock ore drill speed/rebalance fluid use - Multiblock ore drill speed is not very fast. Players would like to speed it up. Should it compare to Ender Quarry? Also, would be interesting to have more fluid options. So, suggestion: ~~- [ ] Reduce fluid use at the drill by 10x. Increase the lube/stone dust use in the mixer by 10x to compensate. This way players can use 256B drums to provide drilling fluid instead of needing bedrockium drums.~~ Players seem ok with this as-is. - [ ] Increase drill base speed. For tier 2, base speed is 2s, adjust up/down appropriately for other tiers. Base speed - need to determine how this applies per tier. Should it all be respect to LV? Or LV = tier 1, MV = tier 2, etc? - [ ] Add API to add custom advanced fluids, and how much faster they are compared to default. IE, fluid.advlube or whatever. Could be 4x faster, maybe add higher tier options. - [ ] Let overclocking dictate base speed adjustments. Let advanced fluid speeds dictate how many blocks are picked per operation. Basic drilling fluid = 1x. Advanced = 4x. Maybe faster ones? - [ ] Add a second fluid hatch for concrete? Let the machine place concrete in the drill holes. Dunno if that matters.
non_process
evaluate multiblock ore drill speed rebalance fluid use multiblock ore drill speed is not very fast players would like to speed it up should it compare to ender quarry also would be interesting to have more fluid options so suggestion reduce fluid use at the drill by increase the lube stone dust use in the mixer by to compensate this way players can use drums to provide drilling fluid instead of needing bedrockium drums players seem ok with this as is increase drill base speed for tier base speed is adjust up down appropriately for other tiers base speed need to determine how this applies per tier should it all be respect to lv or lv tier mv tier etc add api to add custom advanced fluids and how much faster they are compared to default ie fluid advlube or whatever could be faster maybe add higher tier options let overclocking dictate base speed adjustments let advanced fluid speeds dictate how many blocks are picked per operation basic drilling fluid advanced maybe faster ones add a second fluid hatch for concrete let the machine place concrete in the drill holes dunno if that matters
0
9,287
13,100,884,561
IssuesEvent
2020-08-04 01:59:55
goharbor/harbor
https://api.github.com/repos/goharbor/harbor
closed
There are problems with the latest version pushing the same tag image and copy pull command
kind/requirement target/2.1.0 target/FC
**Expected behavior and actual behavior:** 1. How can harbor copy the pull command and get results using sha256 instead of tag.It really affects the experience. ![image](https://user-images.githubusercontent.com/62740686/82034309-9aa3a000-96d0-11ea-89a3-388b6ae9b331.png) 2. I tag the same tag, which means the same version, but now the old version is still there, so there is a lot of garbage data ![image](https://user-images.githubusercontent.com/62740686/82037544-dccee080-96d4-11ea-9bae-41b1828bf1e1.png) **Steps to reproduce the problem:** 1. When you click on the mirror pull command, you get sha256 instead of tag 2. Push the same tag image, check the mirror warehouse after success, and produce a lot of garbage data without tag **Versions:** Please specify the versions of following systems. - harbor version: v2.0.0 - docker engine version: 19.03.8 - docker-compose version: 1.25.5
1.0
There are problems with the latest version pushing the same tag image and copy pull command - **Expected behavior and actual behavior:** 1. How can harbor copy the pull command and get results using sha256 instead of tag.It really affects the experience. ![image](https://user-images.githubusercontent.com/62740686/82034309-9aa3a000-96d0-11ea-89a3-388b6ae9b331.png) 2. I tag the same tag, which means the same version, but now the old version is still there, so there is a lot of garbage data ![image](https://user-images.githubusercontent.com/62740686/82037544-dccee080-96d4-11ea-9bae-41b1828bf1e1.png) **Steps to reproduce the problem:** 1. When you click on the mirror pull command, you get sha256 instead of tag 2. Push the same tag image, check the mirror warehouse after success, and produce a lot of garbage data without tag **Versions:** Please specify the versions of following systems. - harbor version: v2.0.0 - docker engine version: 19.03.8 - docker-compose version: 1.25.5
non_process
there are problems with the latest version pushing the same tag image and copy pull command expected behavior and actual behavior how can harbor copy the pull command and get results using instead of tag it really affects the experience i tag the same tag which means the same version but now the old version is still there so there is a lot of garbage data steps to reproduce the problem when you click on the mirror pull command you get instead of tag push the same tag image check the mirror warehouse after success and produce a lot of garbage data without tag versions please specify the versions of following systems harbor version docker engine version docker compose version
0
48,675
13,392,365,480
IssuesEvent
2020-09-03 01:09:07
jgeraigery/shopping-list-react-pouchdb
https://api.github.com/repos/jgeraigery/shopping-list-react-pouchdb
opened
CVE-2020-7720 (High) detected in node-forge-0.9.0.tgz
security vulnerability
## CVE-2020-7720 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/shopping-list-react-pouchdb/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/shopping-list-react-pouchdb/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.1.tgz (Root Library) - webpack-dev-server-3.10.3.tgz - selfsigned-1.10.7.tgz - :x: **node-forge-0.9.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:3.4.1;webpack-dev-server:3.10.3;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-7720","vulnerabilityDetails":"All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7720 (High) detected in node-forge-0.9.0.tgz - ## CVE-2020-7720 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/shopping-list-react-pouchdb/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/shopping-list-react-pouchdb/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.1.tgz (Root Library) - webpack-dev-server-3.10.3.tgz - selfsigned-1.10.7.tgz - :x: **node-forge-0.9.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:3.4.1;webpack-dev-server:3.10.3;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-7720","vulnerabilityDetails":"All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in node forge tgz cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file tmp ws scm shopping list react pouchdb package json path to vulnerable library tmp ws scm shopping list react pouchdb node modules node forge package json dependency hierarchy react scripts tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library vulnerability details all versions of package node forge are vulnerable to prototype pollution via the util setpath function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails all versions of package node forge are vulnerable to prototype pollution via the util setpath function vulnerabilityurl
0
177,548
14,631,909,130
IssuesEvent
2020-12-23 21:00:37
Josh194/FMCB
https://api.github.com/repos/Josh194/FMCB
closed
Decide on Subsystem General Structure.
documentation in progress
Create first draft for the general structure of the subsystem API. This should not include specific methods, structures, classes, etc, but should lay out the way in which subsystems will access the core buffer libraries, and be built (if applicable) with them.
1.0
Decide on Subsystem General Structure. - Create first draft for the general structure of the subsystem API. This should not include specific methods, structures, classes, etc, but should lay out the way in which subsystems will access the core buffer libraries, and be built (if applicable) with them.
non_process
decide on subsystem general structure create first draft for the general structure of the subsystem api this should not include specific methods structures classes etc but should lay out the way in which subsystems will access the core buffer libraries and be built if applicable with them
0
14,849
18,243,899,692
IssuesEvent
2021-10-01 15:52:05
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
filterprocessor: enable logs filtering by record level attributes
proc: filterprocessor
**Is your feature request related to a problem? Please describe.** This is related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5237. 5237 tries to implement resource level log filtering with regexp whereas this issue aims to address filtering by record level attributes. **Describe the solution you'd like** ~~I have a ready implementation that could be reviewed under https://github.com/pmalek-sumo/opentelemetry-collector-contrib/commit/11de2c03d3dbbde2e73b8a816318a2515c843985 (which would go on top of #5237 when it gets merged)~~ See: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/5279#issuecomment-925705513 **Describe alternatives you've considered** There are no processors that would do it as far as I know.
1.0
filterprocessor: enable logs filtering by record level attributes - **Is your feature request related to a problem? Please describe.** This is related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5237. 5237 tries to implement resource level log filtering with regexp whereas this issue aims to address filtering by record level attributes. **Describe the solution you'd like** ~~I have a ready implementation that could be reviewed under https://github.com/pmalek-sumo/opentelemetry-collector-contrib/commit/11de2c03d3dbbde2e73b8a816318a2515c843985 (which would go on top of #5237 when it gets merged)~~ See: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/5279#issuecomment-925705513 **Describe alternatives you've considered** There are no processors that would do it as far as I know.
process
filterprocessor enable logs filtering by record level attributes is your feature request related to a problem please describe this is related to tries to implement resource level log filtering with regexp whereas this issue aims to address filtering by record level attributes describe the solution you d like i have a ready implementation that could be reviewed under which would go on top of when it gets merged see describe alternatives you ve considered there are no processors that would do it as far as i know
1
9,758
12,741,968,970
IssuesEvent
2020-06-26 07:27:54
commudle/commudle-ng
https://api.github.com/repos/commudle/commudle-ng
closed
Simple online events
advanced api needed brainstorming needed enhancement process flow
Think and make the feature for simple online events. Right now Commudle supports both online and offline events, but it takes a lot of details, including forms to set it up. Let's to make a simple plug and play event feature (the current event setup would be the advanced version of it).
1.0
Simple online events - Think and make the feature for simple online events. Right now Commudle supports both online and offline events, but it takes a lot of details, including forms to set it up. Let's to make a simple plug and play event feature (the current event setup would be the advanced version of it).
process
simple online events think and make the feature for simple online events right now commudle supports both online and offline events but it takes a lot of details including forms to set it up let s to make a simple plug and play event feature the current event setup would be the advanced version of it
1
12,879
15,268,816,220
IssuesEvent
2021-02-22 11:55:48
threefoldfoundation/home
https://api.github.com/repos/threefoldfoundation/home
closed
Marketplace Web - cleanup
process_duplicate
- [ ] now.threefold.io : which is the commercial website about marketplace and has all solutions described (team promotion) - [ ] content from gridsome wiki ex now.threefold.io -> the new now.threefold.io which is based on new web framework (team promotion) - [ ] on each commercial page there is "deploy now button" -> go to auth page, see who user is, redirect to their evdc and their marketplace for that solution - [ ] if you don't exist, ask to install TF Connect
1.0
Marketplace Web - cleanup - - [ ] now.threefold.io : which is the commercial website about marketplace and has all solutions described (team promotion) - [ ] content from gridsome wiki ex now.threefold.io -> the new now.threefold.io which is based on new web framework (team promotion) - [ ] on each commercial page there is "deploy now button" -> go to auth page, see who user is, redirect to their evdc and their marketplace for that solution - [ ] if you don't exist, ask to install TF Connect
process
marketplace web cleanup now threefold io which is the commercial website about marketplace and has all solutions described team promotion content from gridsome wiki ex now threefold io the new now threefold io which is based on new web framework team promotion on each commercial page there is deploy now button go to auth page see who user is redirect to their evdc and their marketplace for that solution if you don t exist ask to install tf connect
1
27,879
6,911,094,577
IssuesEvent
2017-11-28 06:34:29
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Joomla 4 save & close two click procedure.
No Code Attached Yet
### Steps to reproduce the issue open module and save ### Expected result save and close being the default / quick action ### Actual result save is the quick action ### System information (as much as possible) Joomla! 4.0.0-dev Development [ Amani ] 19-November-2017 16:34 GMT ### Additional comments It's completely counterintuitive & counter productive to have to hit a dropdown to save & close. it's far more common that people make all their changes and then save, then it is to make incremental saves. These button choices should be reversed!
1.0
Joomla 4 save & close two click procedure. - ### Steps to reproduce the issue open module and save ### Expected result save and close being the default / quick action ### Actual result save is the quick action ### System information (as much as possible) Joomla! 4.0.0-dev Development [ Amani ] 19-November-2017 16:34 GMT ### Additional comments It's completely counterintuitive & counter productive to have to hit a dropdown to save & close. it's far more common that people make all their changes and then save, then it is to make incremental saves. These button choices should be reversed!
non_process
joomla save close two click procedure steps to reproduce the issue open module and save expected result save and close being the default quick action actual result save is the quick action system information as much as possible joomla dev development november gmt additional comments it s completely counterintuitive counter productive to have to hit a dropdown to save close it s far more common that people make all their changes and then save then it is to make incremental saves these button choices should be reversed
0
5,191
7,973,705,787
IssuesEvent
2018-07-17 00:45:54
bugmark/exchange
https://api.github.com/repos/bugmark/exchange
closed
Use of branches in repo?
9 question DEV PROCESS
It appears that the `master` branch is ahead of the `dev` branch. This is not intuitive for me and I would like clarity of how branches are used in this repo? I ask, because the pull request #82 is against the dev branch, but people will start with the master branch, and if there is no path from dev to master, then I'm confused. Also, if the dev branch is meant for development environment, why should it not have the latest version of features?
1.0
Use of branches in repo? - It appears that the `master` branch is ahead of the `dev` branch. This is not intuitive for me and I would like clarity of how branches are used in this repo? I ask, because the pull request #82 is against the dev branch, but people will start with the master branch, and if there is no path from dev to master, then I'm confused. Also, if the dev branch is meant for development environment, why should it not have the latest version of features?
process
use of branches in repo it appears that the master branch is ahead of the dev branch this is not intuitive for me and i would like clarity of how branches are used in this repo i ask because the pull request is against the dev branch but people will start with the master branch and if there is no path from dev to master then i m confused also if the dev branch is meant for development environment why should it not have the latest version of features
1
8,710
2,611,536,919
IssuesEvent
2015-02-27 06:06:27
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
"The best shot award was won by <b>%1</b> with <b>%2</b> pts." needs plural forms
auto-migrated Priority-Medium Type-Defect
``` "The best shot award was won by <b>%1</b> with <b>%2</b> pts." in PageGameStats needs marking for plural forms, like you already did with "<b>%1</b> thought it's good to shoot his own hedgehogs with <b>%2</b> pts." ``` Original issue reported on code.google.com by `f...@foramnagaidhlig.net` on 4 Aug 2014 at 1:19
1.0
"The best shot award was won by <b>%1</b> with <b>%2</b> pts." needs plural forms - ``` "The best shot award was won by <b>%1</b> with <b>%2</b> pts." in PageGameStats needs marking for plural forms, like you already did with "<b>%1</b> thought it's good to shoot his own hedgehogs with <b>%2</b> pts." ``` Original issue reported on code.google.com by `f...@foramnagaidhlig.net` on 4 Aug 2014 at 1:19
non_process
the best shot award was won by with pts needs plural forms the best shot award was won by with pts in pagegamestats needs marking for plural forms like you already did with thought it s good to shoot his own hedgehogs with pts original issue reported on code google com by f foramnagaidhlig net on aug at
0
2,801
5,731,271,047
IssuesEvent
2017-04-21 11:54:22
g8os/core0
https://api.github.com/repos/g8os/core0
closed
container.bash depends on the $PATH environment variable
process_wontfix type_bug
branch 1.1.0-alpha ```python In [47]: cl.container.create("https://hub.gig.tech/gig-official-apps/ardb-rocksdb.flist", mount={'/mnt/dev-sda1':'/mnt'}, host_network=True) Out[47]: 1 In [50]: container = cl.container.client(1) In [53]: container.bash("ip a").get() Out[53]: STATE: ERROR STDOUT: STDERR: DATA: exec: "sh": executable file not found in $PATH ``` Wouldn't it better to actually use `/bin/bash` ?
1.0
container.bash depends on the $PATH environment variable - branch 1.1.0-alpha ```python In [47]: cl.container.create("https://hub.gig.tech/gig-official-apps/ardb-rocksdb.flist", mount={'/mnt/dev-sda1':'/mnt'}, host_network=True) Out[47]: 1 In [50]: container = cl.container.client(1) In [53]: container.bash("ip a").get() Out[53]: STATE: ERROR STDOUT: STDERR: DATA: exec: "sh": executable file not found in $PATH ``` Wouldn't it better to actually use `/bin/bash` ?
process
container bash depends on the path environment variable branch alpha python in cl container create mount mnt dev mnt host network true out in container cl container client in container bash ip a get out state error stdout stderr data exec sh executable file not found in path wouldn t it better to actually use bin bash
1
3,803
6,782,426,395
IssuesEvent
2017-10-30 08:02:37
decidim/decidim
https://api.github.com/repos/decidim/decidim
closed
Improve process metadata field naming and ordering
component: processes good first issue hacktoberfest team: i18n
# This is a Feature Proposal #### :tophat: Description * Current name for the metadata that is displayed for participatory processes is very confusing. * Example from http://staging.decidim.codegram.com/processes/omnis-sint : ![image](https://user-images.githubusercontent.com/537284/31673088-97e5cf12-b35e-11e7-8af1-15fec926c4b7.png) * What needs to be done is the following: - [ ] change name Participatory scope --> What is decided - [ ] change name Participatory structure --> How is it decided - [ ] change name Target --> Who participates - [ ] local area --> Organization Area - [ ] change order to have it as follows: 1. What is decided, 2. Who participates, 3. How is it decided, 4. Ámbit, 5. Organization Area, 6. Developer Group, 7. Start Date, 8. End date - [ ] Mention @xabier to update the translations :)
1.0
Improve process metadata field naming and ordering - # This is a Feature Proposal #### :tophat: Description * Current name for the metadata that is displayed for participatory processes is very confusing. * Example from http://staging.decidim.codegram.com/processes/omnis-sint : ![image](https://user-images.githubusercontent.com/537284/31673088-97e5cf12-b35e-11e7-8af1-15fec926c4b7.png) * What needs to be done is the following: - [ ] change name Participatory scope --> What is decided - [ ] change name Participatory structure --> How is it decided - [ ] change name Target --> Who participates - [ ] local area --> Organization Area - [ ] change order to have it as follows: 1. What is decided, 2. Who participates, 3. How is it decided, 4. Ámbit, 5. Organization Area, 6. Developer Group, 7. Start Date, 8. End date - [ ] Mention @xabier to update the translations :)
process
improve process metadata field naming and ordering this is a feature proposal tophat description current name for the metadata that is displayed for participatory processes is very confusing example from what needs to be done is the following change name participatory scope what is decided change name participatory structure how is it decided change name target who participates local area organization area change order to have it as follows what is decided who participates how is it decided ámbit organization area developer group start date end date mention xabier to update the translations
1
12,532
14,972,344,762
IssuesEvent
2021-01-27 22:41:01
BootBlock/FileSieve
https://api.github.com/repos/BootBlock/FileSieve
opened
Add a transaction log of all I/O changes made after processing
backend-core enhancement processing
If a detailed-enough log is kept of a full “processing”, then maybe that’ll allow the possibility of “undoing” a processing after-the-fact? Would need proper planning and checks in place as it could get somewhat messy if anything unexpected happens.
1.0
Add a transaction log of all I/O changes made after processing - If a detailed-enough log is kept of a full “processing”, then maybe that’ll allow the possibility of “undoing” a processing after-the-fact? Would need proper planning and checks in place as it could get somewhat messy if anything unexpected happens.
process
add a transaction log of all i o changes made after processing if a detailed enough log is kept of a full “processing” then maybe that’ll allow the possibility of “undoing” a processing after the fact would need proper planning and checks in place as it could get somewhat messy if anything unexpected happens
1
512,488
14,897,962,395
IssuesEvent
2021-01-21 12:30:02
kubermatic/kubermatic
https://api.github.com/repos/kubermatic/kubermatic
closed
Validation: I can create Azure cluster without specifying node size
kind/bug priority/normal sig/ui
I post an azure cluster without specifying node size. Expected that API would validate required fields.
1.0
Validation: I can create Azure cluster without specifying node size - I post an azure cluster without specifying node size. Expected that API would validate required fields.
non_process
validation i can create azure cluster without specifying node size i post an azure cluster without specifying node size expected that api would validate required fields
0
251,225
27,151,001,180
IssuesEvent
2023-02-17 01:14:39
DeepikaKaranji/ContainerApplication-Security-With-JenkinsCI
https://api.github.com/repos/DeepikaKaranji/ContainerApplication-Security-With-JenkinsCI
opened
CVE-2023-23934 (Low) detected in Werkzeug-1.0.1-py2.py3-none-any.whl
security vulnerability
## CVE-2023-23934 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary> <p>The comprehensive WSGI web application library.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p> <p>Path to vulnerable library: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p> <p> Dependency Hierarchy: - Flask-1.1.2-py2.py3-none-any.whl (Root Library) - :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Werkzeug is a comprehensive WSGI web application library. Browsers may allow "nameless" cookies that look like `=value` instead of `key=value`. A vulnerable browser may allow a compromised application on an adjacent subdomain to exploit this to set a cookie like `=__Host-test=bad` for another subdomain. Werkzeug prior to 2.2.3 will parse the cookie `=__Host-test=bad` as __Host-test=bad`. If a Werkzeug application is running next to a vulnerable or malicious subdomain which sets such a cookie using a vulnerable browser, the Werkzeug application will see the bad cookie value but the valid cookie key. The issue is fixed in Werkzeug 2.2.3. <p>Publish Date: 2023-02-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23934>CVE-2023-23934</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23934">https://www.cve.org/CVERecord?id=CVE-2023-23934</a></p> <p>Release Date: 2023-02-14</p> <p>Fix Resolution: Werkzeug - 2.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-23934 (Low) detected in Werkzeug-1.0.1-py2.py3-none-any.whl - ## CVE-2023-23934 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary> <p>The comprehensive WSGI web application library.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p> <p>Path to vulnerable library: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p> <p> Dependency Hierarchy: - Flask-1.1.2-py2.py3-none-any.whl (Root Library) - :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Werkzeug is a comprehensive WSGI web application library. Browsers may allow "nameless" cookies that look like `=value` instead of `key=value`. A vulnerable browser may allow a compromised application on an adjacent subdomain to exploit this to set a cookie like `=__Host-test=bad` for another subdomain. Werkzeug prior to 2.2.3 will parse the cookie `=__Host-test=bad` as __Host-test=bad`. If a Werkzeug application is running next to a vulnerable or malicious subdomain which sets such a cookie using a vulnerable browser, the Werkzeug application will see the bad cookie value but the valid cookie key. The issue is fixed in Werkzeug 2.2.3. <p>Publish Date: 2023-02-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23934>CVE-2023-23934</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23934">https://www.cve.org/CVERecord?id=CVE-2023-23934</a></p> <p>Release Date: 2023-02-14</p> <p>Fix Resolution: Werkzeug - 2.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve low detected in werkzeug none any whl cve low severity vulnerability vulnerable library werkzeug none any whl the comprehensive wsgi web application library library home page a href path to dependency file tmp ws scm container security with jenkinsci requirements txt path to vulnerable library tmp ws scm container security with jenkinsci requirements txt dependency hierarchy flask none any whl root library x werkzeug none any whl vulnerable library vulnerability details werkzeug is a comprehensive wsgi web application library browsers may allow nameless cookies that look like value instead of key value a vulnerable browser may allow a compromised application on an adjacent subdomain to exploit this to set a cookie like host test bad for another subdomain werkzeug prior to will parse the cookie host test bad as host test bad if a werkzeug application is running next to a vulnerable or malicious subdomain which sets such a cookie using a vulnerable browser the werkzeug application will see the bad cookie value but the valid cookie key the issue is fixed in werkzeug publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution werkzeug step up your open source security game with mend
0
520,865
15,096,051,356
IssuesEvent
2021-02-07 13:37:59
RocketMeet/RocketMeet-client
https://api.github.com/repos/RocketMeet/RocketMeet-client
reopened
Comma-separation handler on poll pages not working on mobile
Priority: High bug
From reports by different users, it looks like it could be a browser-related issue too.
1.0
Comma-separation handler on poll pages not working on mobile - From reports by different users, it looks like it could be a browser-related issue too.
non_process
comma separation handler on poll pages not working on mobile from reports by different users it looks like it could be a browser related issue too
0
5,701
8,563,612,108
IssuesEvent
2018-11-09 14:33:26
easy-software-ufal/annotations_repos
https://api.github.com/repos/easy-software-ufal/annotations_repos
opened
aspnet/JsonPatch JsonPatchDocument.Replace() yields invalid path when [JsonProperty] is used (1.1.0)
C# RMP test wrong processing
Issue: `https://github.com/aspnet/JsonPatch/issues/50` PR: `https://github.com/aspnet/JsonPatch/commit/e8452821b9a7689853e36c9a10ad4b8c6281817c`
1.0
aspnet/JsonPatch JsonPatchDocument.Replace() yields invalid path when [JsonProperty] is used (1.1.0) - Issue: `https://github.com/aspnet/JsonPatch/issues/50` PR: `https://github.com/aspnet/JsonPatch/commit/e8452821b9a7689853e36c9a10ad4b8c6281817c`
process
aspnet jsonpatch jsonpatchdocument replace yields invalid path when is used issue pr
1
7,334
10,469,054,019
IssuesEvent
2019-09-22 18:00:02
produvia/ai-platform
https://api.github.com/repos/produvia/ai-platform
closed
Machine Translation
natural-language-processing task wontfix
# Goal(s) - Translate a sentence in a source language to a different target language # Input(s) - Sentence # Output(s) - Sentence # Objective Function(s) - TBD
1.0
Machine Translation - # Goal(s) - Translate a sentence in a source language to a different target language # Input(s) - Sentence # Output(s) - Sentence # Objective Function(s) - TBD
process
machine translation goal s translate a sentence in a source language to a different target language input s sentence output s sentence objective function s tbd
1
209
2,636,660,179
IssuesEvent
2015-03-10 06:39:31
fabric8io/fabric8
https://api.github.com/repos/fabric8io/fabric8
closed
Auto-detect base package of deployed Spring Boot app
java container micro services process manager V1
Now we need to specify the package or base class of the deployed Spring Boot application. For example: profile-edit --pid=io.fabric8.container.java/jvmArguments=-Dspring.main.sources=io.fabric8.process.spring.boot.itests.service.invoicing invoicing I don't like it, because this is error-prone part of the application setup (typo in package name, etc). It will be nice if Fabric8 could try to autodetect base package of the installed jar. This is do-able with some basic jar file scanning. For example: com example project MyClass1.class MyClass2.class Should return `com.example.project`.
1.0
Auto-detect base package of deployed Spring Boot app - Now we need to specify the package or base class of the deployed Spring Boot application. For example: profile-edit --pid=io.fabric8.container.java/jvmArguments=-Dspring.main.sources=io.fabric8.process.spring.boot.itests.service.invoicing invoicing I don't like it, because this is error-prone part of the application setup (typo in package name, etc). It will be nice if Fabric8 could try to autodetect base package of the installed jar. This is do-able with some basic jar file scanning. For example: com example project MyClass1.class MyClass2.class Should return `com.example.project`.
process
auto detect base package of deployed spring boot app now we need to specify the package or base class of the deployed spring boot application for example profile edit pid io container java jvmarguments dspring main sources io process spring boot itests service invoicing invoicing i don t like it because this is error prone part of the application setup typo in package name etc it will be nice if could try to autodetect base package of the installed jar this is do able with some basic jar file scanning for example com example project class class should return com example project
1