Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4,464
| 7,331,139,677
|
IssuesEvent
|
2018-03-05 12:28:28
|
zotero/zotero
|
https://api.github.com/repos/zotero/zotero
|
closed
|
Show "Automatically update citations" in document preferences after document creation
|
Blocker Word Processor Integration
|
People should be able to turn on delayed citations manually after document creation. It shouldn't show up during document creation, but it should show up if you return to the doc prefs.
Discussed here: https://forums.zotero.org/discussion/comment/302684/#Comment_302684
|
1.0
|
Show "Automatically update citations" in document preferences after document creation - People should be able to turn on delayed citations manually after document creation. It shouldn't show up during document creation, but it should show up if you return to the doc prefs.
Discussed here: https://forums.zotero.org/discussion/comment/302684/#Comment_302684
|
process
|
show automatically update citations in document preferences after document creation people should be able to turn on delayed citations manually after document creation it shouldn t show up during document creation but it should show up if you return to the doc prefs discussed here
| 1
|
5,176
| 7,960,135,641
|
IssuesEvent
|
2018-07-13 05:38:11
|
Rokid/ShadowNode
|
https://api.github.com/repos/Rokid/ShadowNode
|
closed
|
child_process: process.send causes memory leaks
|
bug child_process
|
```js
// parent.js
var child = require('child_process').fork(__dirname + '/child.js', {
env: {
isSubprocess: 'true',
}
})
child.on('message', data => {
// console.log(data.toString())
})
//child.js
setInterval(() => {
process.send(Math.random())
}, 0)
```
child's memory increases fast
|
1.0
|
child_process: process.send causes memory leaks - ```js
// parent.js
var child = require('child_process').fork(__dirname + '/child.js', {
env: {
isSubprocess: 'true',
}
})
child.on('message', data => {
// console.log(data.toString())
})
//child.js
setInterval(() => {
process.send(Math.random())
}, 0)
```
child's memory increases fast
|
process
|
child process process send causes memory leaks js parent js var child require child process fork dirname child js env issubprocess true child on message data console log data tostring child js setinterval process send math random child s memory increases fast
| 1
|
10,017
| 13,043,914,329
|
IssuesEvent
|
2020-07-29 03:02:12
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `Uncompress` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `Uncompress` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @sticnarf
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `Uncompress` from TiDB -
## Description
Port the scalar function `Uncompress` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @sticnarf
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function uncompress from tidb description port the scalar function uncompress from tidb to coprocessor score mentor s sticnarf recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
16,836
| 9,536,669,789
|
IssuesEvent
|
2019-04-30 10:20:25
|
Garados007/Werwolf
|
https://api.github.com/repos/Garados007/Werwolf
|
closed
|
Optimiere Abrufe bei Spielrundenwechsel
|
difficult performance
|
Bei einem Wechsel der aktuellen Spielrunde werden ein Großteil der Daten (z.B. Chateinträge) verworfen und müssen neu abgerufen werden. Ein Teil davon ändert sich aber nicht in der nächsten Runde und soll nur ausgeblendet werden oder ungültige registrierte periodische Abrfragen existieren. Diese Abfragen lassen sich optimieren.
|
True
|
Optimiere Abrufe bei Spielrundenwechsel - Bei einem Wechsel der aktuellen Spielrunde werden ein Großteil der Daten (z.B. Chateinträge) verworfen und müssen neu abgerufen werden. Ein Teil davon ändert sich aber nicht in der nächsten Runde und soll nur ausgeblendet werden oder ungültige registrierte periodische Abrfragen existieren. Diese Abfragen lassen sich optimieren.
|
non_process
|
optimiere abrufe bei spielrundenwechsel bei einem wechsel der aktuellen spielrunde werden ein großteil der daten z b chateinträge verworfen und müssen neu abgerufen werden ein teil davon ändert sich aber nicht in der nächsten runde und soll nur ausgeblendet werden oder ungültige registrierte periodische abrfragen existieren diese abfragen lassen sich optimieren
| 0
|
54,772
| 13,920,332,315
|
IssuesEvent
|
2020-10-21 10:17:05
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Cannot convert from UUID to JSON in H2's JSON_OBJECT() and related functions
|
C: DB: H2 C: Functionality E: All Editions P: Medium R: Fixed T: Defect
|
### Expected behavior
Query result containing records with UUID successfully mapped to DTO via JSON document.
### Actual behavior
Exception is thrown when JOOQ attempts to map the record:
```
org.springframework.dao.DataIntegrityViolationException: jOOQ; SQL [select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME"]; Data conversion error converting "UUID to JSON"
Data conversion error converting "UUID to JSON"; SQL statement:
select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]; nested exception is org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON"
Data conversion error converting "UUID to JSON"; SQL statement:
select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]
at org.jooq_3.14.0.DEFAULT.debug(Unknown Source)
at org.springframework.jdbc.support.SQLStateSQLExceptionTranslator.doTranslate(SQLStateSQLExceptionTranslator.java:104)
at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:72)
at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.translate(JooqExceptionTranslator.java:92)
at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.handle(JooqExceptionTranslator.java:81)
at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.exception(JooqExceptionTranslator.java:55)
at org.jooq.impl.ExecuteListeners.exception(ExecuteListeners.java:274)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:390)
at org.jooq.impl.AbstractResultQuery.fetch(AbstractResultQuery.java:333)
at org.jooq.impl.AbstractResultQuery.fetchInto(AbstractResultQuery.java:1545)
at org.jooq.impl.SelectImpl.fetchInto(SelectImpl.java:3936)
at com.example.uuidjson.JooqParentRepository.list(JooqParentRepository.java:38)
at com.example.uuidjson.JooqParentRepository$$FastClassBySpringCGLIB$$93c0a175.invoke(<generated>)
at org.springframework.cglib.proxy.MethodProxy.invoke(MethodProxy.java:218)
at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.invokeJoinpoint(CglibAopProxy.java:771)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:163)
at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749)
at org.springframework.transaction.interceptor.TransactionAspectSupport.invokeWithinTransaction(TransactionAspectSupport.java:367)
at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:118)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:186)
at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749)
at org.springframework.aop.framework.CglibAopProxy$DynamicAdvisedInterceptor.intercept(CglibAopProxy.java:691)
at com.example.uuidjson.JooqParentRepository$$EnhancerBySpringCGLIB$$92b1c668.list(<generated>)
at com.example.uuidjson.JooqParentRepositoryTest.parentWithChildrenFetched(JooqParentRepositoryTest.java:25)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:686)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$6(TestMethodTestDescriptor.java:212)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:208)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:137)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:71)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:135)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1510)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1510)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:32)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:51)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:248)
at org.junit.platform.launcher.core.DefaultLauncher.lambda$execute$5(DefaultLauncher.java:211)
at org.junit.platform.launcher.core.DefaultLauncher.withInterceptedStreams(DefaultLauncher.java:226)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:199)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:132)
at com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:71)
at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33)
at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220)
at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53)
Caused by: org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON"
Data conversion error converting "UUID to JSON"; SQL statement:
select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]
at org.h2.message.DbException.getJdbcSQLException(DbException.java:457)
at org.h2.message.DbException.getJdbcSQLException(DbException.java:429)
at org.h2.message.DbException.get(DbException.java:205)
at org.h2.message.DbException.get(DbException.java:181)
at org.h2.value.Value.getDataConversionError(Value.java:1504)
at org.h2.value.Value.convertToJson(Value.java:1439)
at org.h2.value.Value.convertTo(Value.java:861)
at org.h2.value.Value.convertTo(Value.java:737)
at org.h2.expression.function.Function.jsonObjectAppend(Function.java:2368)
at org.h2.expression.function.Function.jsonObject(Function.java:2351)
at org.h2.expression.function.Function.getSimpleValue(Function.java:1153)
at org.h2.expression.function.Function.getValueWithArgs(Function.java:1274)
at org.h2.expression.function.Function.getValue(Function.java:672)
at org.h2.expression.aggregate.Aggregate.updateAggregate(Aggregate.java:205)
at org.h2.expression.aggregate.AbstractAggregate.updateAggregate(AbstractAggregate.java:280)
at org.h2.expression.analysis.DataAnalysisOperation.updateAggregate(DataAnalysisOperation.java:252)
at org.h2.expression.function.Function.updateAggregate(Function.java:3049)
at org.h2.command.dml.Select.updateAgg(Select.java:545)
at org.h2.command.dml.Select.gatherGroup(Select.java:525)
at org.h2.command.dml.Select.queryGroup(Select.java:487)
at org.h2.command.dml.Select.queryWithoutCache(Select.java:839)
at org.h2.command.dml.Query.queryWithoutCacheLazyCheck(Query.java:201)
at org.h2.command.dml.Query.query(Query.java:489)
at org.h2.command.dml.Query.query(Query.java:451)
at org.h2.command.CommandContainer.query(CommandContainer.java:285)
at org.h2.command.Command.executeQuery(Command.java:195)
at org.h2.jdbc.JdbcPreparedStatement.execute(JdbcPreparedStatement.java:233)
at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:214)
at org.jooq.impl.Tools.executeStatementAndGetFirstResultSet(Tools.java:4186)
at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:279)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:375)
... 79 more
```
### Steps to reproduce the problem
1. Create table with UUID as primary key:
```
create table parent(
id UUID constraint pk_parent_id primary key,
name varchar(100)
);
```
2. Run code generation, check that generated field is of UUID type:
```
public final TableField<ParentRecord, UUID> ID = createField(DSL.name("ID"), SQLDataType.UUID.nullable(false), this, "");
```
3. Create target DTO for record mapper:
```
@Getter
@RequiredArgsConstructor
public class Parent {
private final UUID id;
private final String name;
private final Collection<Child> children;
}
```
4. Run a query mapping its result into DTO using JSON as described here: https://blog.jooq.org/2020/10/09/nesting-collections-with-jooq-3-14s-sql-xml-or-sql-json-support/
```
dsl.select(
jsonObject(
jsonEntry("id", PARENT.ID),
jsonEntry("name", PARENT.NAME),
jsonEntry("children", jsonArrayAgg(
jsonObject(
jsonEntry("parentId", CHILD.PARENT_ID),
jsonEntry("name", CHILD.NAME)
)
).orderBy(PARENT.NAME))
)
)
.from(PARENT)
.leftJoin(CHILD).on(CHILD.PARENT_ID.eq(PARENT.ID))
.groupBy(PARENT.ID, PARENT.NAME)
.orderBy(PARENT.ID, PARENT.NAME)
.fetchInto(Parent.class);
```
### Versions
- jOOQ: 3.14.0
- Java: 11
- Database: h2
- OS: Windows
- JDBC Driver: org.h2.Driver
|
1.0
|
Cannot convert from UUID to JSON in H2's JSON_OBJECT() and related functions - ### Expected behavior
Query result containing records with UUID successfully mapped to DTO via JSON document.
### Actual behavior
Exception is thrown when JOOQ attempts to map the record:
```
org.springframework.dao.DataIntegrityViolationException: jOOQ; SQL [select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME"]; Data conversion error converting "UUID to JSON"
Data conversion error converting "UUID to JSON"; SQL statement:
select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]; nested exception is org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON"
Data conversion error converting "UUID to JSON"; SQL statement:
select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]
at org.jooq_3.14.0.DEFAULT.debug(Unknown Source)
at org.springframework.jdbc.support.SQLStateSQLExceptionTranslator.doTranslate(SQLStateSQLExceptionTranslator.java:104)
at org.springframework.jdbc.support.AbstractFallbackSQLExceptionTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:72)
at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.translate(JooqExceptionTranslator.java:92)
at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.handle(JooqExceptionTranslator.java:81)
at org.springframework.boot.autoconfigure.jooq.JooqExceptionTranslator.exception(JooqExceptionTranslator.java:55)
at org.jooq.impl.ExecuteListeners.exception(ExecuteListeners.java:274)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:390)
at org.jooq.impl.AbstractResultQuery.fetch(AbstractResultQuery.java:333)
at org.jooq.impl.AbstractResultQuery.fetchInto(AbstractResultQuery.java:1545)
at org.jooq.impl.SelectImpl.fetchInto(SelectImpl.java:3936)
at com.example.uuidjson.JooqParentRepository.list(JooqParentRepository.java:38)
at com.example.uuidjson.JooqParentRepository$$FastClassBySpringCGLIB$$93c0a175.invoke(<generated>)
at org.springframework.cglib.proxy.MethodProxy.invoke(MethodProxy.java:218)
at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.invokeJoinpoint(CglibAopProxy.java:771)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:163)
at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749)
at org.springframework.transaction.interceptor.TransactionAspectSupport.invokeWithinTransaction(TransactionAspectSupport.java:367)
at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:118)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:186)
at org.springframework.aop.framework.CglibAopProxy$CglibMethodInvocation.proceed(CglibAopProxy.java:749)
at org.springframework.aop.framework.CglibAopProxy$DynamicAdvisedInterceptor.intercept(CglibAopProxy.java:691)
at com.example.uuidjson.JooqParentRepository$$EnhancerBySpringCGLIB$$92b1c668.list(<generated>)
at com.example.uuidjson.JooqParentRepositoryTest.parentWithChildrenFetched(JooqParentRepositoryTest.java:25)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:686)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$6(TestMethodTestDescriptor.java:212)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:208)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:137)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:71)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:135)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1510)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1510)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:38)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$5(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$7(NodeTestTask.java:125)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:135)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:122)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:80)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:32)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:51)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:248)
at org.junit.platform.launcher.core.DefaultLauncher.lambda$execute$5(DefaultLauncher.java:211)
at org.junit.platform.launcher.core.DefaultLauncher.withInterceptedStreams(DefaultLauncher.java:226)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:199)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:132)
at com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:71)
at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33)
at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220)
at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53)
Caused by: org.h2.jdbc.JdbcSQLDataException: Data conversion error converting "UUID to JSON"
Data conversion error converting "UUID to JSON"; SQL statement:
select json_object(key ? value "PARENT"."ID", key ? value "PARENT"."NAME", key ? value json_arrayagg(json_object(key ? value "CHILD"."PARENT_ID", key ? value "CHILD"."NAME") order by "PARENT"."NAME")) from "PARENT" left outer join "CHILD" on "CHILD"."PARENT_ID" = "PARENT"."ID" group by "PARENT"."ID", "PARENT"."NAME" order by "PARENT"."ID", "PARENT"."NAME" [22018-200]
at org.h2.message.DbException.getJdbcSQLException(DbException.java:457)
at org.h2.message.DbException.getJdbcSQLException(DbException.java:429)
at org.h2.message.DbException.get(DbException.java:205)
at org.h2.message.DbException.get(DbException.java:181)
at org.h2.value.Value.getDataConversionError(Value.java:1504)
at org.h2.value.Value.convertToJson(Value.java:1439)
at org.h2.value.Value.convertTo(Value.java:861)
at org.h2.value.Value.convertTo(Value.java:737)
at org.h2.expression.function.Function.jsonObjectAppend(Function.java:2368)
at org.h2.expression.function.Function.jsonObject(Function.java:2351)
at org.h2.expression.function.Function.getSimpleValue(Function.java:1153)
at org.h2.expression.function.Function.getValueWithArgs(Function.java:1274)
at org.h2.expression.function.Function.getValue(Function.java:672)
at org.h2.expression.aggregate.Aggregate.updateAggregate(Aggregate.java:205)
at org.h2.expression.aggregate.AbstractAggregate.updateAggregate(AbstractAggregate.java:280)
at org.h2.expression.analysis.DataAnalysisOperation.updateAggregate(DataAnalysisOperation.java:252)
at org.h2.expression.function.Function.updateAggregate(Function.java:3049)
at org.h2.command.dml.Select.updateAgg(Select.java:545)
at org.h2.command.dml.Select.gatherGroup(Select.java:525)
at org.h2.command.dml.Select.queryGroup(Select.java:487)
at org.h2.command.dml.Select.queryWithoutCache(Select.java:839)
at org.h2.command.dml.Query.queryWithoutCacheLazyCheck(Query.java:201)
at org.h2.command.dml.Query.query(Query.java:489)
at org.h2.command.dml.Query.query(Query.java:451)
at org.h2.command.CommandContainer.query(CommandContainer.java:285)
at org.h2.command.Command.executeQuery(Command.java:195)
at org.h2.jdbc.JdbcPreparedStatement.execute(JdbcPreparedStatement.java:233)
at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:214)
at org.jooq.impl.Tools.executeStatementAndGetFirstResultSet(Tools.java:4186)
at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:279)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:375)
... 79 more
```
### Steps to reproduce the problem
1. Create table with UUID as primary key:
```
create table parent(
id UUID constraint pk_parent_id primary key,
name varchar(100)
);
```
2. Run code generation, check that generated field is of UUID type:
```
public final TableField<ParentRecord, UUID> ID = createField(DSL.name("ID"), SQLDataType.UUID.nullable(false), this, "");
```
3. Create target DTO for record mapper:
```
@Getter
@RequiredArgsConstructor
public class Parent {
private final UUID id;
private final String name;
private final Collection<Child> children;
}
```
4. Run a query mapping its result into DTO using JSON as described here: https://blog.jooq.org/2020/10/09/nesting-collections-with-jooq-3-14s-sql-xml-or-sql-json-support/
```
dsl.select(
jsonObject(
jsonEntry("id", PARENT.ID),
jsonEntry("name", PARENT.NAME),
jsonEntry("children", jsonArrayAgg(
jsonObject(
jsonEntry("parentId", CHILD.PARENT_ID),
jsonEntry("name", CHILD.NAME)
)
).orderBy(PARENT.NAME))
)
)
.from(PARENT)
.leftJoin(CHILD).on(CHILD.PARENT_ID.eq(PARENT.ID))
.groupBy(PARENT.ID, PARENT.NAME)
.orderBy(PARENT.ID, PARENT.NAME)
.fetchInto(Parent.class);
```
### Versions
- jOOQ: 3.14.0
- Java: 11
- Database: h2
- OS: Windows
- JDBC Driver: org.h2.Driver
|
non_process
|
cannot convert from uuid to json in s json object and related functions expected behavior query result containing records with uuid successfully mapped to dto via json document actual behavior exception is thrown when jooq attempts to map the record org springframework dao dataintegrityviolationexception jooq sql data conversion error converting uuid to json data conversion error converting uuid to json sql statement select json object key value parent id key value parent name key value json arrayagg json object key value child parent id key value child name order by parent name from parent left outer join child on child parent id parent id group by parent id parent name order by parent id parent name nested exception is org jdbc jdbcsqldataexception data conversion error converting uuid to json data conversion error converting uuid to json sql statement select json object key value parent id key value parent name key value json arrayagg json object key value child parent id key value child name order by parent name from parent left outer join child on child parent id parent id group by parent id parent name order by parent id parent name at org jooq default debug unknown source at org springframework jdbc support sqlstatesqlexceptiontranslator dotranslate sqlstatesqlexceptiontranslator java at org springframework jdbc support abstractfallbacksqlexceptiontranslator translate abstractfallbacksqlexceptiontranslator java at org springframework boot autoconfigure jooq jooqexceptiontranslator translate jooqexceptiontranslator java at org springframework boot autoconfigure jooq jooqexceptiontranslator handle jooqexceptiontranslator java at org springframework boot autoconfigure jooq jooqexceptiontranslator exception jooqexceptiontranslator java at org jooq impl executelisteners exception executelisteners java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractresultquery fetch abstractresultquery java at org jooq impl abstractresultquery fetchinto abstractresultquery java at org jooq impl selectimpl fetchinto selectimpl java at com example uuidjson jooqparentrepository list jooqparentrepository java at com example uuidjson jooqparentrepository fastclassbyspringcglib invoke at org springframework cglib proxy methodproxy invoke methodproxy java at org springframework aop framework cglibaopproxy cglibmethodinvocation invokejoinpoint cglibaopproxy java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop framework cglibaopproxy cglibmethodinvocation proceed cglibaopproxy java at org springframework transaction interceptor transactionaspectsupport invokewithintransaction transactionaspectsupport java at org springframework transaction interceptor transactioninterceptor invoke transactioninterceptor java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop framework cglibaopproxy cglibmethodinvocation proceed cglibaopproxy java at org springframework aop framework cglibaopproxy dynamicadvisedinterceptor intercept cglibaopproxy java at com example uuidjson jooqparentrepository enhancerbyspringcglib list at com example uuidjson jooqparentrepositorytest parentwithchildrenfetched jooqparentrepositorytest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit platform commons util reflectionutils invokemethod reflectionutils java at org junit jupiter engine execution methodinvocation proceed methodinvocation java at org junit jupiter engine execution invocationinterceptorchain validatinginvocation proceed invocationinterceptorchain java at org junit jupiter engine extension timeoutextension intercept timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestablemethod timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestmethod timeoutextension java at org junit jupiter engine execution executableinvoker reflectiveinterceptorcall lambda ofvoidmethod executableinvoker java at org junit jupiter engine execution executableinvoker lambda invoke executableinvoker java at org junit jupiter engine execution invocationinterceptorchain interceptedinvocation proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain chainandinvoke invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain invoke invocationinterceptorchain java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine descriptor testmethodtestdescriptor lambda invoketestmethod testmethodtestdescriptor java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit jupiter engine descriptor testmethodtestdescriptor invoketestmethod testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at java base java util arraylist foreach arraylist java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice invokeall samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at java base java util arraylist foreach arraylist java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice invokeall samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice submit samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical hierarchicaltestexecutor execute hierarchicaltestexecutor java at org junit platform engine support hierarchical hierarchicaltestengine execute hierarchicaltestengine java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher lambda execute defaultlauncher java at org junit platform launcher core defaultlauncher withinterceptedstreams defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at com intellij startrunnerwithargs java at com intellij rt junit ideatestrunner repeater startrunnerwithargs ideatestrunner java at com intellij rt junit junitstarter preparestreamsandstart junitstarter java at com intellij rt junit junitstarter main junitstarter java caused by org jdbc jdbcsqldataexception data conversion error converting uuid to json data conversion error converting uuid to json sql statement select json object key value parent id key value parent name key value json arrayagg json object key value child parent id key value child name order by parent name from parent left outer join child on child parent id parent id group by parent id parent name order by parent id parent name at org message dbexception getjdbcsqlexception dbexception java at org message dbexception getjdbcsqlexception dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org value value getdataconversionerror value java at org value value converttojson value java at org value value convertto value java at org value value convertto value java at org expression function function jsonobjectappend function java at org expression function function jsonobject function java at org expression function function getsimplevalue function java at org expression function function getvaluewithargs function java at org expression function function getvalue function java at org expression aggregate aggregate updateaggregate aggregate java at org expression aggregate abstractaggregate updateaggregate abstractaggregate java at org expression analysis dataanalysisoperation updateaggregate dataanalysisoperation java at org expression function function updateaggregate function java at org command dml select updateagg select java at org command dml select gathergroup select java at org command dml select querygroup select java at org command dml select querywithoutcache select java at org command dml query querywithoutcachelazycheck query java at org command dml query query query java at org command dml query query query java at org command commandcontainer query commandcontainer java at org command command executequery command java at org jdbc jdbcpreparedstatement execute jdbcpreparedstatement java at org jooq tools jdbc defaultpreparedstatement execute defaultpreparedstatement java at org jooq impl tools executestatementandgetfirstresultset tools java at org jooq impl abstractresultquery execute abstractresultquery java at org jooq impl abstractquery execute abstractquery java more steps to reproduce the problem create table with uuid as primary key create table parent id uuid constraint pk parent id primary key name varchar run code generation check that generated field is of uuid type public final tablefield id createfield dsl name id sqldatatype uuid nullable false this create target dto for record mapper getter requiredargsconstructor public class parent private final uuid id private final string name private final collection children run a query mapping its result into dto using json as described here dsl select jsonobject jsonentry id parent id jsonentry name parent name jsonentry children jsonarrayagg jsonobject jsonentry parentid child parent id jsonentry name child name orderby parent name from parent leftjoin child on child parent id eq parent id groupby parent id parent name orderby parent id parent name fetchinto parent class versions jooq java database os windows jdbc driver org driver
| 0
|
17,812
| 23,739,991,520
|
IssuesEvent
|
2022-08-31 11:34:59
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[dialog] 文档或示例中缺少open-type-event的调用案例
|
good first issue in process
|
### 这个功能解决了什么问题
dialog的确认键如果想使用button的开放能力获取手机号,用哪个字段、如何声明getPhoneNumber呢
### 你建议的方案是什么
示例中给个使用的案例吧
|
1.0
|
[dialog] 文档或示例中缺少open-type-event的调用案例 - ### 这个功能解决了什么问题
dialog的确认键如果想使用button的开放能力获取手机号,用哪个字段、如何声明getPhoneNumber呢
### 你建议的方案是什么
示例中给个使用的案例吧
|
process
|
文档或示例中缺少open type event的调用案例 这个功能解决了什么问题 dialog的确认键如果想使用button的开放能力获取手机号,用哪个字段、如何声明getphonenumber呢 你建议的方案是什么 示例中给个使用的案例吧
| 1
|
105,725
| 9,100,269,255
|
IssuesEvent
|
2019-02-20 08:01:42
|
humera987/FXLabs-Test-Automation
|
https://api.github.com/repos/humera987/FXLabs-Test-Automation
|
opened
|
Test : ApiV1OrgsIdPutOrgorgplanbasicUsercDisallowAbact2
|
test
|
Project : Test
Job : Default
Env : Default
Category : null
Tags : null
Severity : null
Region : US_WEST
Result : fail
Status Code : 500
Headers : {}
Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/
Request :
{
"billingEmail" : "katrina.reynolds@yahoo.com",
"company" : "Kuvalis-Kuvalis",
"createdBy" : "",
"createdDate" : "",
"description" : "cXD8m8jM",
"id" : "",
"inactive" : false,
"location" : "cXD8m8jM",
"modifiedBy" : "",
"modifiedDate" : "",
"name" : "cXD8m8jM",
"orgPlan" : "TEAM",
"orgType" : "TEAM",
"version" : ""
}
Response :
I/O error on PUT request for "http://13.56.210.25/api/v1/api/v1/orgs/": Read timed out; nested exception is java.net.SocketTimeoutException: Read timed out
Logs :
Assertion [@StatusCode == 401 OR @StatusCode == 403 OR @Response.errors == true] resolved-to [500 == 401 OR 500 == 403 OR == true] result [Failed]
--- FX Bot ---
|
1.0
|
Test : ApiV1OrgsIdPutOrgorgplanbasicUsercDisallowAbact2 - Project : Test
Job : Default
Env : Default
Category : null
Tags : null
Severity : null
Region : US_WEST
Result : fail
Status Code : 500
Headers : {}
Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/
Request :
{
"billingEmail" : "katrina.reynolds@yahoo.com",
"company" : "Kuvalis-Kuvalis",
"createdBy" : "",
"createdDate" : "",
"description" : "cXD8m8jM",
"id" : "",
"inactive" : false,
"location" : "cXD8m8jM",
"modifiedBy" : "",
"modifiedDate" : "",
"name" : "cXD8m8jM",
"orgPlan" : "TEAM",
"orgType" : "TEAM",
"version" : ""
}
Response :
I/O error on PUT request for "http://13.56.210.25/api/v1/api/v1/orgs/": Read timed out; nested exception is java.net.SocketTimeoutException: Read timed out
Logs :
Assertion [@StatusCode == 401 OR @StatusCode == 403 OR @Response.errors == true] resolved-to [500 == 401 OR 500 == 403 OR == true] result [Failed]
--- FX Bot ---
|
non_process
|
test project test job default env default category null tags null severity null region us west result fail status code headers endpoint request billingemail katrina reynolds yahoo com company kuvalis kuvalis createdby createddate description id inactive false location modifiedby modifieddate name orgplan team orgtype team version response i o error on put request for read timed out nested exception is java net sockettimeoutexception read timed out logs assertion resolved to result fx bot
| 0
|
77,675
| 3,507,208,456
|
IssuesEvent
|
2016-01-08 11:54:51
|
OregonCore/OregonCore
|
https://api.github.com/repos/OregonCore/OregonCore
|
closed
|
Crash Alert (BB #655)
|
Category: Crash migrated Priority: Medium Type: Bug
|
This issue was migrated from bitbucket.
**Original Reporter:** PadreWoW
**Original Date:** 20.08.2014 09:22:23 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** closed
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/655
<hr>
2014-08-20 13:20:33 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:39 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:44 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:49 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:57 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:21:05 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
|
1.0
|
Crash Alert (BB #655) - This issue was migrated from bitbucket.
**Original Reporter:** PadreWoW
**Original Date:** 20.08.2014 09:22:23 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** closed
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/655
<hr>
2014-08-20 13:20:33 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:39 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:44 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:49 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:20:57 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
2014-08-20 13:21:05 CRASH ALERT: Unit (Entry 12922) is trying to delete its updating MG (Type 5)!
|
non_process
|
crash alert bb this issue was migrated from bitbucket original reporter padrewow original date gmt original priority major original type bug original state closed direct link crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type crash alert unit entry is trying to delete its updating mg type
| 0
|
453,359
| 13,068,578,819
|
IssuesEvent
|
2020-07-31 04:01:52
|
ProjectSidewalk/SidewalkWebpage
|
https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage
|
closed
|
In Version 2.1, No Smooth Animated Interpolation Between Locations When Double Clicking
|
Priority: Low wontfix
|
In version 2.1, there is no smooth animated interpolation between locations when double clicking to move. See below. Is there some function parameter we have to pass to enable this smooth movement animation on the double click to move? It's enabled when you click the arrow to move.

|
1.0
|
In Version 2.1, No Smooth Animated Interpolation Between Locations When Double Clicking - In version 2.1, there is no smooth animated interpolation between locations when double clicking to move. See below. Is there some function parameter we have to pass to enable this smooth movement animation on the double click to move? It's enabled when you click the arrow to move.

|
non_process
|
in version no smooth animated interpolation between locations when double clicking in version there is no smooth animated interpolation between locations when double clicking to move see below is there some function parameter we have to pass to enable this smooth movement animation on the double click to move it s enabled when you click the arrow to move
| 0
|
4,638
| 7,482,326,321
|
IssuesEvent
|
2018-04-05 00:39:08
|
UnbFeelings/unb-feelings-GQA
|
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
|
closed
|
Mudar o template da descrição de atividades
|
process wiki
|
Mudar o template atual que é o seguinte:
<table>
<tbody>
<tr>
<td><b>Descrição</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Papéis</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Entradas</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Saídas</b></td>
<td>Texto</td>
</tr>
</tbody>
</table>
Para:
<table>
<tbody>
<tr>
<td><b>Descrição</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Responsáveis</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Envolvidos</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Entradas</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Saídas</b></td>
<td>Texto</td>
</tr>
</tbody>
</table>
Resumindo: Na descrição das atividades,colocar :
- Responsáveis
- Envolvidos
Em substituição aos papéis
|
1.0
|
Mudar o template da descrição de atividades - Mudar o template atual que é o seguinte:
<table>
<tbody>
<tr>
<td><b>Descrição</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Papéis</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Entradas</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Saídas</b></td>
<td>Texto</td>
</tr>
</tbody>
</table>
Para:
<table>
<tbody>
<tr>
<td><b>Descrição</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Responsáveis</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Envolvidos</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Entradas</b></td>
<td>Texto</td>
</tr>
<tr>
<td><b>Saídas</b></td>
<td>Texto</td>
</tr>
</tbody>
</table>
Resumindo: Na descrição das atividades,colocar :
- Responsáveis
- Envolvidos
Em substituição aos papéis
|
process
|
mudar o template da descrição de atividades mudar o template atual que é o seguinte descrição texto papéis texto entradas texto saídas texto para descrição texto responsáveis texto envolvidos texto entradas texto saídas texto resumindo na descrição das atividades colocar responsáveis envolvidos em substituição aos papéis
| 1
|
137,153
| 12,746,763,540
|
IssuesEvent
|
2020-06-26 16:32:57
|
Ameelio/letters-api
|
https://api.github.com/repos/Ameelio/letters-api
|
closed
|
Add example Request and example Response bodies to User endpoints
|
documentation
|
The following endpoints need example Request and/or Response bodies in the API Documentation Wiki page.
GET /api/user
GET /api/user/{id}
GET /api/users/contacts
GET /api/users/letters
GET /api/users/org
|
1.0
|
Add example Request and example Response bodies to User endpoints - The following endpoints need example Request and/or Response bodies in the API Documentation Wiki page.
GET /api/user
GET /api/user/{id}
GET /api/users/contacts
GET /api/users/letters
GET /api/users/org
|
non_process
|
add example request and example response bodies to user endpoints the following endpoints need example request and or response bodies in the api documentation wiki page get api user get api user id get api users contacts get api users letters get api users org
| 0
|
22,229
| 30,780,346,815
|
IssuesEvent
|
2023-07-31 09:34:44
|
EBIvariation/eva-opentargets
|
https://api.github.com/repos/EBIvariation/eva-opentargets
|
opened
|
Evidence string generation for 2023.09 release
|
Processing
|
**Deadline for submission: 8 August 2023**
Refer to [documentation](https://github.com/EBIvariation/eva-opentargets/blob/master/docs/generate-evidence-strings.md) for full description of steps.
|
1.0
|
Evidence string generation for 2023.09 release - **Deadline for submission: 8 August 2023**
Refer to [documentation](https://github.com/EBIvariation/eva-opentargets/blob/master/docs/generate-evidence-strings.md) for full description of steps.
|
process
|
evidence string generation for release deadline for submission august refer to for full description of steps
| 1
|
770,751
| 27,054,804,634
|
IssuesEvent
|
2023-02-13 15:29:58
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.zeit.de - site is not usable
|
priority-important browser-firefox-tablet engine-gecko
|
<!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 4.2.2; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/118150 -->
**URL**: https://www.zeit.de/zustimmung?url=https%3A%2F%2Fwww.zeit.de%2Fentdecken%2F2023-02%2Fjudith-holofernes-wochenende-tipps-podcast
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android 4.2.2
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
I can't choice the button 'Einverstanden und weiter mit Werbung'
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/2/20e0bea9-62d6-421a-9d17-6532d11934ff.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827153727</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/2/bc3bc6fc-b10b-4d72-a83a-edd635629f05)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.zeit.de - site is not usable - <!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 4.2.2; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/118150 -->
**URL**: https://www.zeit.de/zustimmung?url=https%3A%2F%2Fwww.zeit.de%2Fentdecken%2F2023-02%2Fjudith-holofernes-wochenende-tipps-podcast
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android 4.2.2
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
I can't choice the button 'Einverstanden und weiter mit Werbung'
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/2/20e0bea9-62d6-421a-9d17-6532d11934ff.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827153727</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/2/bc3bc6fc-b10b-4d72-a83a-edd635629f05)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
site is not usable url browser version firefox mobile tablet operating system android tested another browser yes chrome problem type site is not usable description missing items steps to reproduce i can t choice the button einverstanden und weiter mit werbung view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
780,618
| 27,401,754,735
|
IssuesEvent
|
2023-03-01 01:34:29
|
hotosm/fmtm
|
https://api.github.com/repos/hotosm/fmtm
|
closed
|
Add ODK->OSM conversion
|
enhancement Priority: Must have in progress
|
Use the odkconvert project conversion program that takes an CSV submission files and converts it to good OSM XML.
|
1.0
|
Add ODK->OSM conversion - Use the odkconvert project conversion program that takes an CSV submission files and converts it to good OSM XML.
|
non_process
|
add odk osm conversion use the odkconvert project conversion program that takes an csv submission files and converts it to good osm xml
| 0
|
20,880
| 31,466,909,721
|
IssuesEvent
|
2023-08-30 03:16:35
|
BG3-Community-Library-Team/BG3-Community-Library
|
https://api.github.com/repos/BG3-Community-Library-Team/BG3-Community-Library
|
closed
|
Bladesinger Subclass Compatibility Support
|
Subclass Compatibilty Framework
|
**Is your feature request related to a problem? Please describe.**
We now have permission to add support for the [Bladesinger](https://www.nexusmods.com/baldursgate3/mods/279) subclass.
**Describe the solution you'd like**
It will need to be implemented in the same way as [PR 39](https://github.com/BG3-Community-Library-Team/BG3-Community-Library/pull/39), which handled some fighter subclasses.
|
True
|
Bladesinger Subclass Compatibility Support - **Is your feature request related to a problem? Please describe.**
We now have permission to add support for the [Bladesinger](https://www.nexusmods.com/baldursgate3/mods/279) subclass.
**Describe the solution you'd like**
It will need to be implemented in the same way as [PR 39](https://github.com/BG3-Community-Library-Team/BG3-Community-Library/pull/39), which handled some fighter subclasses.
|
non_process
|
bladesinger subclass compatibility support is your feature request related to a problem please describe we now have permission to add support for the subclass describe the solution you d like it will need to be implemented in the same way as which handled some fighter subclasses
| 0
|
9,794
| 12,807,791,993
|
IssuesEvent
|
2020-07-03 12:17:12
|
solid/process
|
https://api.github.com/repos/solid/process
|
reopened
|
Improving the effectiveness of panels
|
process issue
|
We [originally introduced](https://github.com/solid/process/pull/24) panels as part of the Solid process ten months ago. Our aim was to drive focused and thoughtful collaboration around specific topics, leading to meaningful contributions to the Solid Specification and ecosystem. We held off on adding a lot of structure around how panels are formed, or how they operate, until we had some real experiences _in the wild_ to learn from.
Work towards a normative [Solid Specification](https://github.com/solid/specification) has picked up considerably. We just merged the first [candidate proposal](https://github.com/solid/specification/pull/156) from a panel into the editors draft. We have some real experiences to draw from now, and a lot of work left to be done. It's time to look at how panels have operated to date, and consider any opportunities to improve them. How can we help them drive positive, constructive contributions to the Solid specification and ecosystem in a timely manner?
Here are a few considerations and thoughts based on personal observations and discussions with panelists, editors, and others:
- There are too many panels, and only a few of them are active and/or productive. We should have less panels with greater coordination and more focus.
- There's no clear or consistent organizing structure in active panels. Who is responsible to set expectations, hold or be held accountable, and keep focus where it needs to be?
- Active panels aren't focusing enough on the most pressing needs of the specification, or the most important use cases in the real-world.
- The editorial team needs to provide more active support to panels, and better communication around specification priorities.
_This list is not meant to be exhaustive, only to contribute to the discussion in this thread._
Whatever your role (panelist, community member, editor, etc) - please chime in and share your thoughts on how you'd like to see panels evolve.
|
1.0
|
Improving the effectiveness of panels -
We [originally introduced](https://github.com/solid/process/pull/24) panels as part of the Solid process ten months ago. Our aim was to drive focused and thoughtful collaboration around specific topics, leading to meaningful contributions to the Solid Specification and ecosystem. We held off on adding a lot of structure around how panels are formed, or how they operate, until we had some real experiences _in the wild_ to learn from.
Work towards a normative [Solid Specification](https://github.com/solid/specification) has picked up considerably. We just merged the first [candidate proposal](https://github.com/solid/specification/pull/156) from a panel into the editors draft. We have some real experiences to draw from now, and a lot of work left to be done. It's time to look at how panels have operated to date, and consider any opportunities to improve them. How can we help them drive positive, constructive contributions to the Solid specification and ecosystem in a timely manner?
Here are a few considerations and thoughts based on personal observations and discussions with panelists, editors, and others:
- There are too many panels, and only a few of them are active and/or productive. We should have less panels with greater coordination and more focus.
- There's no clear or consistent organizing structure in active panels. Who is responsible to set expectations, hold or be held accountable, and keep focus where it needs to be?
- Active panels aren't focusing enough on the most pressing needs of the specification, or the most important use cases in the real-world.
- The editorial team needs to provide more active support to panels, and better communication around specification priorities.
_This list is not meant to be exhaustive, only to contribute to the discussion in this thread._
Whatever your role (panelist, community member, editor, etc) - please chime in and share your thoughts on how you'd like to see panels evolve.
|
process
|
improving the effectiveness of panels we panels as part of the solid process ten months ago our aim was to drive focused and thoughtful collaboration around specific topics leading to meaningful contributions to the solid specification and ecosystem we held off on adding a lot of structure around how panels are formed or how they operate until we had some real experiences in the wild to learn from work towards a normative has picked up considerably we just merged the first from a panel into the editors draft we have some real experiences to draw from now and a lot of work left to be done it s time to look at how panels have operated to date and consider any opportunities to improve them how can we help them drive positive constructive contributions to the solid specification and ecosystem in a timely manner here are a few considerations and thoughts based on personal observations and discussions with panelists editors and others there are too many panels and only a few of them are active and or productive we should have less panels with greater coordination and more focus there s no clear or consistent organizing structure in active panels who is responsible to set expectations hold or be held accountable and keep focus where it needs to be active panels aren t focusing enough on the most pressing needs of the specification or the most important use cases in the real world the editorial team needs to provide more active support to panels and better communication around specification priorities this list is not meant to be exhaustive only to contribute to the discussion in this thread whatever your role panelist community member editor etc please chime in and share your thoughts on how you d like to see panels evolve
| 1
|
460,844
| 13,219,005,217
|
IssuesEvent
|
2020-08-17 09:42:58
|
MyDataTaiwan/logboard
|
https://api.github.com/repos/MyDataTaiwan/logboard
|
opened
|
When user switch data window by "Today", "This Week" .. buttons, the calendar on top should also sync
|
QA priority-high
|
**Steps to Reproduce**
1. User upload data to LogBoard
2. User switch the data window by the buttons "Today", "This Week"...... etc.
**Results**




* Expected:
When user switch data windows, the date in the top calendar should also sync with the data window
* Actual:
The date in the top calendar won't alter when switching data window by buttons
**Logs**
* <upload-error-log-here>
**Environment**
* mylog14-dashboard: v0.5.2-beta5
|
1.0
|
When user switch data window by "Today", "This Week" .. buttons, the calendar on top should also sync - **Steps to Reproduce**
1. User upload data to LogBoard
2. User switch the data window by the buttons "Today", "This Week"...... etc.
**Results**




* Expected:
When user switch data windows, the date in the top calendar should also sync with the data window
* Actual:
The date in the top calendar won't alter when switching data window by buttons
**Logs**
* <upload-error-log-here>
**Environment**
* mylog14-dashboard: v0.5.2-beta5
|
non_process
|
when user switch data window by today this week buttons the calendar on top should also sync steps to reproduce user upload data to logboard user switch the data window by the buttons today this week etc results expected when user switch data windows the date in the top calendar should also sync with the data window actual the date in the top calendar won t alter when switching data window by buttons logs environment dashboard
| 0
|
20,379
| 2,622,845,847
|
IssuesEvent
|
2015-03-04 08:02:31
|
max99x/pagemon-chrome-ext
|
https://api.github.com/repos/max99x/pagemon-chrome-ext
|
closed
|
Switch to the new favicon API
|
auto-migrated Priority-Low Type-Enhancement
|
```
Chrome 5+ exposes chrome://favicon/ URLs that can be used to get favicons by
simply supplying a domain name. No longer do we need to explicitly save
favicon URLs.
```
Original issue reported on code.google.com by `max99x` on 6 Jun 2010 at 3:31
|
1.0
|
Switch to the new favicon API - ```
Chrome 5+ exposes chrome://favicon/ URLs that can be used to get favicons by
simply supplying a domain name. No longer do we need to explicitly save
favicon URLs.
```
Original issue reported on code.google.com by `max99x` on 6 Jun 2010 at 3:31
|
non_process
|
switch to the new favicon api chrome exposes chrome favicon urls that can be used to get favicons by simply supplying a domain name no longer do we need to explicitly save favicon urls original issue reported on code google com by on jun at
| 0
|
20,958
| 27,817,251,759
|
IssuesEvent
|
2023-03-18 20:37:14
|
cse442-at-ub/project_s23-cinco
|
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
|
closed
|
Add routing to profile, account settings, and login button on navbar
|
Processing Task Sprint 2
|
*Task Tests*
run in "Sprint2-Navbar-Buttons" branch in github
test1:
- in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser
- look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"
-in a mobile view, click on the burger icon to ensure the following buttons are included: "profile, "login", "account settings":
-click on "profile" to get to edit profile page:
-click on "account settings" to get to the account settings page:
-click on "login" to get to login page:
|
1.0
|
Add routing to profile, account settings, and login button on navbar - *Task Tests*
run in "Sprint2-Navbar-Buttons" branch in github
test1:
- in the project folder (project_s23-cinco) run "npm start" in the terminal, this will open the homepage in a browser
- look at the navbar and ensure it has the following buttons: "profile", "account settings", "login"
-in a mobile view, click on the burger icon to ensure the following buttons are included: "profile, "login", "account settings":
-click on "profile" to get to edit profile page:
-click on "account settings" to get to the account settings page:
-click on "login" to get to login page:
|
process
|
add routing to profile account settings and login button on navbar task tests run in navbar buttons branch in github in the project folder project cinco run npm start in the terminal this will open the homepage in a browser look at the navbar and ensure it has the following buttons profile account settings login in a mobile view click on the burger icon to ensure the following buttons are included profile login account settings click on profile to get to edit profile page click on account settings to get to the account settings page click on login to get to login page
| 1
|
4,623
| 4,489,147,864
|
IssuesEvent
|
2016-08-30 09:55:32
|
syl20bnr/spacemacs
|
https://api.github.com/repos/syl20bnr/spacemacs
|
closed
|
Spaceline performance bug?
|
- Bug tracker - Fixed upstream Performance
|
#### Description
with Emacs 25.1.50.3 (built from source, git commit 8cfd9ba…, with -O2 and gtk toolkit on Linux Mint 17.3, using the spacemacs develop branch and freshly updated packages as of May 1st 2016), I experience extremely laggy cursor motion. Just writing a single line of text into an otherwise empty text-mode buffer and scrolling back and forth causes the cursor to stutter every ~20 characters.
Disabling the modeline via t-m-t eliminates the issue; with spaceline enabled, I notice that I get on the order of 5 garbage collections for every line traversal (roughly 100 characters).
Memory profiling shows that spaceline indeed allocates a surprisingly large amount of memory (this profile only shows the cost of 1 or 2 line traversals, `command-execute` down below was caused by helm when I toggled the profiler via `helm-M-x-read-extended-command`):
```
- redisplay_internal (C function) 19,812,999 70%
- eval 19,812,999 70%
- spaceline--prepare 19,350,659 68%
- spaceline--prepare-any 18,402,191 65%
- spaceline--eval-segment 14,209,815 50%
+ mapcar 6,775,993 24%
+ powerline-raw 1,064,653 3%
+ spaceline--eval-segment 320,340 1%
+ spaceline--segment-major-mode 127,352 0%
+ eval 5,024 0%
+ mapcar 1,389,412 4%
+ spaceline--get-face 565,504 2%
+ powerline-render 593,944 2%
+ powerline-width 349,444 1%
powerline-fill 5,080 0%
+ command-execute 7,656,897 27%
+ winner-save-old-configurations 686,565 2%
+ evil-repeat-pre-hook 8,188 0%
sp--save-pre-command-state 2,112 0%
+ command-error-default-function 2,032 0%
internal-timer-start-idle 1,746 0%
+ timer-event-handler 1,572 0%
... 0 0%
```
For now, I'm working around the issue by setting the garbage collection threshold to 1GB and triggering collection after 5 idle seconds:
```
(setq gc-cons-threshold 1000000000)
(run-with-idle-timer 5 t #'garbage-collect)
```
While this has made basic text navigation bearable again, I still find it a bit unsettling to know that spaceline is doing a lot of probably unnecessary work in the background. Could it be worthwhile to change spaceline from its current, nicely-functional style to a perhaps more efficient imperative style? Or is this performance problem completely unexpected and possibly caused by some configuration error on my side?
#### Reproduction guide
- Start Emacs
- Move the cursor around
#### System Info
- OS: gnu/linux
- Emacs: 25.1.50.3
- Spacemacs: 0.105.19
- Spacemacs branch: c-improved (rev. 1e67a11) (based on develop, with minor changes to c/c++-mode packages.el)
- Graphic display: t
- Distribution: spacemacs
- Editing style: vim
- Completion: helm
- Layers:
```elisp
(spacemacs-helm auto-completion
(c-c++ :variables c-c++-enable-clang-support t c-c++-enable-irony-support t)
colors context emacs-lisp ess evil-snipe
(git :variables git-magit-status-fullscreen t)
github gtags haskell html
(latex :variables latex-enable-auto-fill nil)
lua java javascript markdown org python ranger ruby
(shell :variables shell-default-height 30 shell-default-position 'bottom)
syntax-checking themes-megapack typescript typography
(version-control :variables version-control-diff-tool 'git-gutter)
(vinegar :variables vinegar-reuse-dired-buffer t)
ycmd)
```
|
True
|
Spaceline performance bug? - #### Description
with Emacs 25.1.50.3 (built from source, git commit 8cfd9ba…, with -O2 and gtk toolkit on Linux Mint 17.3, using the spacemacs develop branch and freshly updated packages as of May 1st 2016), I experience extremely laggy cursor motion. Just writing a single line of text into an otherwise empty text-mode buffer and scrolling back and forth causes the cursor to stutter every ~20 characters.
Disabling the modeline via t-m-t eliminates the issue; with spaceline enabled, I notice that I get on the order of 5 garbage collections for every line traversal (roughly 100 characters).
Memory profiling shows that spaceline indeed allocates a surprisingly large amount of memory (this profile only shows the cost of 1 or 2 line traversals, `command-execute` down below was caused by helm when I toggled the profiler via `helm-M-x-read-extended-command`):
```
- redisplay_internal (C function) 19,812,999 70%
- eval 19,812,999 70%
- spaceline--prepare 19,350,659 68%
- spaceline--prepare-any 18,402,191 65%
- spaceline--eval-segment 14,209,815 50%
+ mapcar 6,775,993 24%
+ powerline-raw 1,064,653 3%
+ spaceline--eval-segment 320,340 1%
+ spaceline--segment-major-mode 127,352 0%
+ eval 5,024 0%
+ mapcar 1,389,412 4%
+ spaceline--get-face 565,504 2%
+ powerline-render 593,944 2%
+ powerline-width 349,444 1%
powerline-fill 5,080 0%
+ command-execute 7,656,897 27%
+ winner-save-old-configurations 686,565 2%
+ evil-repeat-pre-hook 8,188 0%
sp--save-pre-command-state 2,112 0%
+ command-error-default-function 2,032 0%
internal-timer-start-idle 1,746 0%
+ timer-event-handler 1,572 0%
... 0 0%
```
For now, I'm working around the issue by setting the garbage collection threshold to 1GB and triggering collection after 5 idle seconds:
```
(setq gc-cons-threshold 1000000000)
(run-with-idle-timer 5 t #'garbage-collect)
```
While this has made basic text navigation bearable again, I still find it a bit unsettling to know that spaceline is doing a lot of probably unnecessary work in the background. Could it be worthwhile to change spaceline from its current, nicely-functional style to a perhaps more efficient imperative style? Or is this performance problem completely unexpected and possibly caused by some configuration error on my side?
#### Reproduction guide
- Start Emacs
- Move the cursor around
#### System Info
- OS: gnu/linux
- Emacs: 25.1.50.3
- Spacemacs: 0.105.19
- Spacemacs branch: c-improved (rev. 1e67a11) (based on develop, with minor changes to c/c++-mode packages.el)
- Graphic display: t
- Distribution: spacemacs
- Editing style: vim
- Completion: helm
- Layers:
```elisp
(spacemacs-helm auto-completion
(c-c++ :variables c-c++-enable-clang-support t c-c++-enable-irony-support t)
colors context emacs-lisp ess evil-snipe
(git :variables git-magit-status-fullscreen t)
github gtags haskell html
(latex :variables latex-enable-auto-fill nil)
lua java javascript markdown org python ranger ruby
(shell :variables shell-default-height 30 shell-default-position 'bottom)
syntax-checking themes-megapack typescript typography
(version-control :variables version-control-diff-tool 'git-gutter)
(vinegar :variables vinegar-reuse-dired-buffer t)
ycmd)
```
|
non_process
|
spaceline performance bug description with emacs built from source git commit … with and gtk toolkit on linux mint using the spacemacs develop branch and freshly updated packages as of may i experience extremely laggy cursor motion just writing a single line of text into an otherwise empty text mode buffer and scrolling back and forth causes the cursor to stutter every characters disabling the modeline via t m t eliminates the issue with spaceline enabled i notice that i get on the order of garbage collections for every line traversal roughly characters memory profiling shows that spaceline indeed allocates a surprisingly large amount of memory this profile only shows the cost of or line traversals command execute down below was caused by helm when i toggled the profiler via helm m x read extended command redisplay internal c function eval spaceline prepare spaceline prepare any spaceline eval segment mapcar powerline raw spaceline eval segment spaceline segment major mode eval mapcar spaceline get face powerline render powerline width powerline fill command execute winner save old configurations evil repeat pre hook sp save pre command state command error default function internal timer start idle timer event handler for now i m working around the issue by setting the garbage collection threshold to and triggering collection after idle seconds setq gc cons threshold run with idle timer t garbage collect while this has made basic text navigation bearable again i still find it a bit unsettling to know that spaceline is doing a lot of probably unnecessary work in the background could it be worthwhile to change spaceline from its current nicely functional style to a perhaps more efficient imperative style or is this performance problem completely unexpected and possibly caused by some configuration error on my side reproduction guide start emacs move the cursor around system info os gnu linux emacs spacemacs spacemacs branch c improved rev based on develop with minor changes to c c mode packages el graphic display t distribution spacemacs editing style vim completion helm layers elisp spacemacs helm auto completion c c variables c c enable clang support t c c enable irony support t colors context emacs lisp ess evil snipe git variables git magit status fullscreen t github gtags haskell html latex variables latex enable auto fill nil lua java javascript markdown org python ranger ruby shell variables shell default height shell default position bottom syntax checking themes megapack typescript typography version control variables version control diff tool git gutter vinegar variables vinegar reuse dired buffer t ycmd
| 0
|
195,159
| 14,705,589,952
|
IssuesEvent
|
2021-01-04 18:23:34
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Number of secrets rewritten during encryption process is seen incorrect
|
[zube]: To Test kind/bug-qa
|
**What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Deploy a cluster, and create some secrets
- K get secrets -A --> On the cluster gives 137 secrets
- enable secrets encryption
```
services:
kube-api:
secrets_encryption_config:
enabled: true
```
- View in API of the cluster --> Do a rotateEncryptionKey action.
- Logs of rancher server show that 150 secrets have been rewritten.
- Logs in Rancher server:
```
2020/12/17 04:28:54 [INFO] cluster [c-fzvn2] provisioning: Rewriting cluster secrets
2020/12/17 04:29:02 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 50 secrets rewritten
2020/12/17 04:29:14 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 100 secrets rewritten
2020/12/17 04:29:26 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 150 secrets rewritten
2020/12/17 04:29:29 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] Operation completed
```
**Expected Result:**
The exact secrets rewritten must be shown. (maybe it is rounded off to the next batch size)
**Other details that may be helpful:**
Related to the fix for - https://github.com/rancher/rancher/issues/27735
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.5-head - commit id: `bb69b03`
- Installation option (single install/HA): Single node
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): RKE DO
- Kubernetes version (use `kubectl version`):
```
1.19 (Any)
```
|
1.0
|
Number of secrets rewritten during encryption process is seen incorrect - **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Deploy a cluster, and create some secrets
- K get secrets -A --> On the cluster gives 137 secrets
- enable secrets encryption
```
services:
kube-api:
secrets_encryption_config:
enabled: true
```
- View in API of the cluster --> Do a rotateEncryptionKey action.
- Logs of rancher server show that 150 secrets have been rewritten.
- Logs in Rancher server:
```
2020/12/17 04:28:54 [INFO] cluster [c-fzvn2] provisioning: Rewriting cluster secrets
2020/12/17 04:29:02 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 50 secrets rewritten
2020/12/17 04:29:14 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 100 secrets rewritten
2020/12/17 04:29:26 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] 150 secrets rewritten
2020/12/17 04:29:29 [INFO] cluster [c-fzvn2] provisioning: [rewrite-secrets] Operation completed
```
**Expected Result:**
The exact secrets rewritten must be shown. (maybe it is rounded off to the next batch size)
**Other details that may be helpful:**
Related to the fix for - https://github.com/rancher/rancher/issues/27735
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.5-head - commit id: `bb69b03`
- Installation option (single install/HA): Single node
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): RKE DO
- Kubernetes version (use `kubectl version`):
```
1.19 (Any)
```
|
non_process
|
number of secrets rewritten during encryption process is seen incorrect what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible deploy a cluster and create some secrets k get secrets a on the cluster gives secrets enable secrets encryption services kube api secrets encryption config enabled true view in api of the cluster do a rotateencryptionkey action logs of rancher server show that secrets have been rewritten logs in rancher server cluster provisioning rewriting cluster secrets cluster provisioning secrets rewritten cluster provisioning secrets rewritten cluster provisioning secrets rewritten cluster provisioning operation completed expected result the exact secrets rewritten must be shown maybe it is rounded off to the next batch size other details that may be helpful related to the fix for environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui head commit id installation option single install ha single node if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported rke do kubernetes version use kubectl version any
| 0
|
284,719
| 21,466,688,488
|
IssuesEvent
|
2022-04-26 05:02:22
|
appsmithorg/appsmith
|
https://api.github.com/repos/appsmithorg/appsmith
|
closed
|
[Docs] #13254 [Bug]-[800]:Close button can be clicked on a disabled select widget
|
Documentation User Education Pod
|
> TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | App Viewers Pod
**Parent Ticket** | #13254
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
1.0
|
[Docs] #13254 [Bug]-[800]:Close button can be clicked on a disabled select widget - > TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | App Viewers Pod
**Parent Ticket** | #13254
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
non_process
|
close button can be clicked on a disabled select widget todo evaluate if this task is needed if not add the skip docs label on the parent ticket fill these fields prepare first draft add label ready for docs team field details pod app viewers pod parent ticket engineer release date live date first draft auto assign priority environment
| 0
|
136,102
| 30,475,790,514
|
IssuesEvent
|
2023-07-17 16:24:23
|
ita-social-projects/StreetCode
|
https://api.github.com/repos/ita-social-projects/StreetCode
|
opened
|
[Admin/History map] The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block
|
bug (Epic#2) Admin/New StreetCode
|
**Environment:** OS: Windows 10 Pro
Browser: Google Chrome Version 111.0.5563.112.
**Reproducible:** always.
**Build found:** commit [d494c37](https://github.com/ita-social-projects/StreetCode/commit/d494c372c230bf30fef322fdb50405a1c708c55b)
**Type:** Functional
**Priority:** Low
**Severity:** Low
**Preconditions**
1. Go to the site.
2. Login as admin.
3. Open the new StreetCode page or the StreetCode page for editing.
**Steps to reproduce:**
1. Go to the "Мапа історії" block.
2. Enter the address of location the streetcode to the "Додати стріткод на мапу" form or click on the desired location on the map.
3. Enter the number of the streetcode plate with a negative value.
4. Pay atention to the information about the location of the streetcode under the map.
**Actual result**
The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block.
**Expected result**
The system restricts adding the number of the streetcode plate with a negative value.
**User story and test case links**
User story #126
Test case #362
<img width="1125" alt="number" src="https://github.com/ita-social-projects/StreetCode/assets/135837034/6578f8ec-35fe-4837-a740-5aabb318b228">
|
1.0
|
[Admin/History map] The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block - **Environment:** OS: Windows 10 Pro
Browser: Google Chrome Version 111.0.5563.112.
**Reproducible:** always.
**Build found:** commit [d494c37](https://github.com/ita-social-projects/StreetCode/commit/d494c372c230bf30fef322fdb50405a1c708c55b)
**Type:** Functional
**Priority:** Low
**Severity:** Low
**Preconditions**
1. Go to the site.
2. Login as admin.
3. Open the new StreetCode page or the StreetCode page for editing.
**Steps to reproduce:**
1. Go to the "Мапа історії" block.
2. Enter the address of location the streetcode to the "Додати стріткод на мапу" form or click on the desired location on the map.
3. Enter the number of the streetcode plate with a negative value.
4. Pay atention to the information about the location of the streetcode under the map.
**Actual result**
The number of the streetcode plate with a negative value is added to the map in the "Мапа історії" block.
**Expected result**
The system restricts adding the number of the streetcode plate with a negative value.
**User story and test case links**
User story #126
Test case #362
<img width="1125" alt="number" src="https://github.com/ita-social-projects/StreetCode/assets/135837034/6578f8ec-35fe-4837-a740-5aabb318b228">
|
non_process
|
the number of the streetcode plate with a negative value is added to the map in the мапа історії block environment os windows pro browser google chrome version reproducible always build found commit type functional priority low severity low preconditions go to the site login as admin open the new streetcode page or the streetcode page for editing steps to reproduce go to the мапа історії block enter the address of location the streetcode to the додати стріткод на мапу form or click on the desired location on the map enter the number of the streetcode plate with a negative value pay atention to the information about the location of the streetcode under the map actual result the number of the streetcode plate with a negative value is added to the map in the мапа історії block expected result the system restricts adding the number of the streetcode plate with a negative value user story and test case links user story test case img width alt number src
| 0
|
16,784
| 21,970,817,600
|
IssuesEvent
|
2022-05-25 03:25:18
|
quark-engine/quark-engine
|
https://api.github.com/repos/quark-engine/quark-engine
|
closed
|
Update author information
|
issue-processing-state-06
|
Since the main author was changed, we should update the author information in setup.py.
|
1.0
|
Update author information - Since the main author was changed, we should update the author information in setup.py.
|
process
|
update author information since the main author was changed we should update the author information in setup py
| 1
|
3,589
| 6,621,672,102
|
IssuesEvent
|
2017-09-21 20:06:36
|
WikiWatershed/rapid-watershed-delineation
|
https://api.github.com/repos/WikiWatershed/rapid-watershed-delineation
|
closed
|
Expose unsimplified result
|
BigCZ Geoprocessing API
|
Provide a way for users to get the unsimplified geojson results. The MMW Geoprocessing API would like to expose them.
|
1.0
|
Expose unsimplified result - Provide a way for users to get the unsimplified geojson results. The MMW Geoprocessing API would like to expose them.
|
process
|
expose unsimplified result provide a way for users to get the unsimplified geojson results the mmw geoprocessing api would like to expose them
| 1
|
19,591
| 25,932,202,326
|
IssuesEvent
|
2022-12-16 10:59:19
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Node hangs when running a stand-alone script with captured output
|
help wanted windows process
|
* **Version**: v8.1.3
* **Platform**: Windows 10 version 10.0.14393 x64
* **Subsystem**:
I am running into an issue on our build servers where Node hangs after running a stand-alone script if it is launched with its standard handles redirected to pipes. Specifically, I am using Ruby's `popen3` function, which wraps OS pipe functionality to obtain pipe streams and then passes the underlying handles to the C runtime library's `popen` API function. I don't _believe_ there is anything particular about the way Ruby is doing this, and I have the expectation that similar code written in C++ directly calling the underlying API functions would exhibit the same problem. I have not actually tested this, though.
I first observed this behaviour as part of an `npm install`. When NPM installs the `node-sass` module, it tries to run an install hook script, and the resulting `node scripts/install.js` process hangs until it is sent a break signal or terminated externally. This does not occur if `npm install` is simply run at the command-line, nor does it occur with simple piping done via the command interpreter (`npm install | more`, for instance).
In further debugging, I tried installing a local copy of `node-sass` in which I made alterations to the `scripts/install.js` script. The hang is reproducible even if the entire `install.js` script is a single `console.log` line. Furthermore, if I outright remove the `scripts/install.js` file, then the parent process of what would have been `node scripts/install.js` encounters an error (expected), complains loudly about it, and then hangs. The command-line of this parent process is:
C:\Program Files\nodejs\\node.exe "C:\NPM\node_modules\npm\bin\npm-cli.js" install "C:\Path\To\node-sass"
I'm not sure what the next step is in identifying the root cause of this issue. I suspect NPM might play some part in it, because if I try my test on a dummy script file _directly_ (`node test.js`), then the hang is not observed.
The following Ruby code should demonstrate the hang if the `node-sass` module is configured to be installed into `node_modules` in the current project directory:
require 'open3'
Open3.popen3("npm install node-sass") do |stdin, stdout, stderr, thr|
stdout_thread = Thread.new { stdout.each_line { |line| puts line } }
stderr_thread = Thread.new { stderr.each_line { |line| puts "[STDERR] #{line}" } }
stdout_thread.join
stderr_thread.join
puts "Exit code: #{thr.value.exitstatus}"
end
At the time of the hang, if I attach a debugger to the rogue `node.exe` process, I see the following threads:
Main thread:
=> Blocked in OS API: GetQueuedCompletionStatusEx
=> Call stack:
node.exe!uv_poll_ex(uv_loop_s * loop, unsigned long timeout) Line 432 C
node.exe!uv_run(uv_loop_s * loop, uv_run_mode mode) Line 513 C
node.exe!node::Start(v8::Isolate * isolate, node::IsolateData * isolate_data, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4520 C++
node.exe!node::Start(uv_loop_s * event_loop, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4585 C++
node.exe!node::Start(int argc, char * * argv) Line 4640 C++
[Inline Frame] node.exe!invoke_main() Line 79 C++
Worker threads: 8 threads are in this state
=> Blocked in OS API: WaitForSingleObjectEx
=> Call stack:
[Inline Frame] node.exe!v8::base::Semaphore::Wait() Line 159 C++
[Inline Frame] node.exe!v8::platform::TaskQueue::GetNext() Line 46 C++
node.exe!v8::platform::WorkerThread::Run() Line 25 C++
[Inline Frame] node.exe!v8::base::Thread::NotifyStartedAndRun() Line 512 C++
node.exe!v8::base::ThreadEntry(void * arg) Line 1328 C++
[Inline Frame] node.exe!invoke_thread_procedure(unsigned int(*)(void *) context, void * const) Line 91 C++
node.exe!thread_start<unsigned int (__cdecl*)(void * __ptr64)>(void * const parameter) Line 115 C++
Thread pool threads: 4 threads are in this state
=> Blocked in OS API: NtWaitForWorkViaWorkerFactory
=> Call stack:
ntdll.dll!TppWorkerThread()
Note that the process that is hanging is _not_ the direct descendant of the calling site that is waiting (in this case the Ruby interpreter). The child process launched by `popen3` has in fact exited, but the _streams_ (`stdour`, `stderr`) aren't signalled because they were inherited by grandchildren, and one of those is the culprit that isn't exiting.
[WAITING] Ruby code: `popen "npm install"`
`- [exited] cmd.exe npm.bat
|- [exited] node.exe npm-cli.js prefix -g
`- [exited] node.exe npm-cli.js install
`- [exited] cmd.exe /d /s /c node.exe scripts/install.js
`- [STILL RUNNING/HUNG] node.exe scripts/install.js
The hung process is sitting around keeping `stdout` and `stderr` open, and since the pattern of the call site is to wait for the end of those streams as the indicator that the child process tree is complete, the call site blocks indefinitely as well.
I have experimented with altering the call site so that it instead explicitly waits for its direct child process to exit, and forces the streams closed. This does permit the calling code to proceed, but the rogue process still remains, which (since this is Windows) causes issues later on to do with file locking.
If there is further debugging I can do to try to narrow down the cause of this issue, I would appreciate any advice. :-) Thanks!
|
1.0
|
Node hangs when running a stand-alone script with captured output - * **Version**: v8.1.3
* **Platform**: Windows 10 version 10.0.14393 x64
* **Subsystem**:
I am running into an issue on our build servers where Node hangs after running a stand-alone script if it is launched with its standard handles redirected to pipes. Specifically, I am using Ruby's `popen3` function, which wraps OS pipe functionality to obtain pipe streams and then passes the underlying handles to the C runtime library's `popen` API function. I don't _believe_ there is anything particular about the way Ruby is doing this, and I have the expectation that similar code written in C++ directly calling the underlying API functions would exhibit the same problem. I have not actually tested this, though.
I first observed this behaviour as part of an `npm install`. When NPM installs the `node-sass` module, it tries to run an install hook script, and the resulting `node scripts/install.js` process hangs until it is sent a break signal or terminated externally. This does not occur if `npm install` is simply run at the command-line, nor does it occur with simple piping done via the command interpreter (`npm install | more`, for instance).
In further debugging, I tried installing a local copy of `node-sass` in which I made alterations to the `scripts/install.js` script. The hang is reproducible even if the entire `install.js` script is a single `console.log` line. Furthermore, if I outright remove the `scripts/install.js` file, then the parent process of what would have been `node scripts/install.js` encounters an error (expected), complains loudly about it, and then hangs. The command-line of this parent process is:
C:\Program Files\nodejs\\node.exe "C:\NPM\node_modules\npm\bin\npm-cli.js" install "C:\Path\To\node-sass"
I'm not sure what the next step is in identifying the root cause of this issue. I suspect NPM might play some part in it, because if I try my test on a dummy script file _directly_ (`node test.js`), then the hang is not observed.
The following Ruby code should demonstrate the hang if the `node-sass` module is configured to be installed into `node_modules` in the current project directory:
require 'open3'
Open3.popen3("npm install node-sass") do |stdin, stdout, stderr, thr|
stdout_thread = Thread.new { stdout.each_line { |line| puts line } }
stderr_thread = Thread.new { stderr.each_line { |line| puts "[STDERR] #{line}" } }
stdout_thread.join
stderr_thread.join
puts "Exit code: #{thr.value.exitstatus}"
end
At the time of the hang, if I attach a debugger to the rogue `node.exe` process, I see the following threads:
Main thread:
=> Blocked in OS API: GetQueuedCompletionStatusEx
=> Call stack:
node.exe!uv_poll_ex(uv_loop_s * loop, unsigned long timeout) Line 432 C
node.exe!uv_run(uv_loop_s * loop, uv_run_mode mode) Line 513 C
node.exe!node::Start(v8::Isolate * isolate, node::IsolateData * isolate_data, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4520 C++
node.exe!node::Start(uv_loop_s * event_loop, int argc, const char * const * argv, int exec_argc, const char * const * exec_argv) Line 4585 C++
node.exe!node::Start(int argc, char * * argv) Line 4640 C++
[Inline Frame] node.exe!invoke_main() Line 79 C++
Worker threads: 8 threads are in this state
=> Blocked in OS API: WaitForSingleObjectEx
=> Call stack:
[Inline Frame] node.exe!v8::base::Semaphore::Wait() Line 159 C++
[Inline Frame] node.exe!v8::platform::TaskQueue::GetNext() Line 46 C++
node.exe!v8::platform::WorkerThread::Run() Line 25 C++
[Inline Frame] node.exe!v8::base::Thread::NotifyStartedAndRun() Line 512 C++
node.exe!v8::base::ThreadEntry(void * arg) Line 1328 C++
[Inline Frame] node.exe!invoke_thread_procedure(unsigned int(*)(void *) context, void * const) Line 91 C++
node.exe!thread_start<unsigned int (__cdecl*)(void * __ptr64)>(void * const parameter) Line 115 C++
Thread pool threads: 4 threads are in this state
=> Blocked in OS API: NtWaitForWorkViaWorkerFactory
=> Call stack:
ntdll.dll!TppWorkerThread()
Note that the process that is hanging is _not_ the direct descendant of the calling site that is waiting (in this case the Ruby interpreter). The child process launched by `popen3` has in fact exited, but the _streams_ (`stdour`, `stderr`) aren't signalled because they were inherited by grandchildren, and one of those is the culprit that isn't exiting.
[WAITING] Ruby code: `popen "npm install"`
`- [exited] cmd.exe npm.bat
|- [exited] node.exe npm-cli.js prefix -g
`- [exited] node.exe npm-cli.js install
`- [exited] cmd.exe /d /s /c node.exe scripts/install.js
`- [STILL RUNNING/HUNG] node.exe scripts/install.js
The hung process is sitting around keeping `stdout` and `stderr` open, and since the pattern of the call site is to wait for the end of those streams as the indicator that the child process tree is complete, the call site blocks indefinitely as well.
I have experimented with altering the call site so that it instead explicitly waits for its direct child process to exit, and forces the streams closed. This does permit the calling code to proceed, but the rogue process still remains, which (since this is Windows) causes issues later on to do with file locking.
If there is further debugging I can do to try to narrow down the cause of this issue, I would appreciate any advice. :-) Thanks!
|
process
|
node hangs when running a stand alone script with captured output version platform windows version subsystem i am running into an issue on our build servers where node hangs after running a stand alone script if it is launched with its standard handles redirected to pipes specifically i am using ruby s function which wraps os pipe functionality to obtain pipe streams and then passes the underlying handles to the c runtime library s popen api function i don t believe there is anything particular about the way ruby is doing this and i have the expectation that similar code written in c directly calling the underlying api functions would exhibit the same problem i have not actually tested this though i first observed this behaviour as part of an npm install when npm installs the node sass module it tries to run an install hook script and the resulting node scripts install js process hangs until it is sent a break signal or terminated externally this does not occur if npm install is simply run at the command line nor does it occur with simple piping done via the command interpreter npm install more for instance in further debugging i tried installing a local copy of node sass in which i made alterations to the scripts install js script the hang is reproducible even if the entire install js script is a single console log line furthermore if i outright remove the scripts install js file then the parent process of what would have been node scripts install js encounters an error expected complains loudly about it and then hangs the command line of this parent process is c program files nodejs node exe c npm node modules npm bin npm cli js install c path to node sass i m not sure what the next step is in identifying the root cause of this issue i suspect npm might play some part in it because if i try my test on a dummy script file directly node test js then the hang is not observed the following ruby code should demonstrate the hang if the node sass module is configured to be installed into node modules in the current project directory require npm install node sass do stdin stdout stderr thr stdout thread thread new stdout each line line puts line stderr thread thread new stderr each line line puts line stdout thread join stderr thread join puts exit code thr value exitstatus end at the time of the hang if i attach a debugger to the rogue node exe process i see the following threads main thread blocked in os api getqueuedcompletionstatusex call stack node exe uv poll ex uv loop s loop unsigned long timeout line c node exe uv run uv loop s loop uv run mode mode line c node exe node start isolate isolate node isolatedata isolate data int argc const char const argv int exec argc const char const exec argv line c node exe node start uv loop s event loop int argc const char const argv int exec argc const char const exec argv line c node exe node start int argc char argv line c node exe invoke main line c worker threads threads are in this state blocked in os api waitforsingleobjectex call stack node exe base semaphore wait line c node exe platform taskqueue getnext line c node exe platform workerthread run line c node exe base thread notifystartedandrun line c node exe base threadentry void arg line c node exe invoke thread procedure unsigned int void context void const line c node exe thread start void const parameter line c thread pool threads threads are in this state blocked in os api ntwaitforworkviaworkerfactory call stack ntdll dll tppworkerthread note that the process that is hanging is not the direct descendant of the calling site that is waiting in this case the ruby interpreter the child process launched by has in fact exited but the streams stdour stderr aren t signalled because they were inherited by grandchildren and one of those is the culprit that isn t exiting ruby code popen npm install cmd exe npm bat node exe npm cli js prefix g node exe npm cli js install cmd exe d s c node exe scripts install js node exe scripts install js the hung process is sitting around keeping stdout and stderr open and since the pattern of the call site is to wait for the end of those streams as the indicator that the child process tree is complete the call site blocks indefinitely as well i have experimented with altering the call site so that it instead explicitly waits for its direct child process to exit and forces the streams closed this does permit the calling code to proceed but the rogue process still remains which since this is windows causes issues later on to do with file locking if there is further debugging i can do to try to narrow down the cause of this issue i would appreciate any advice thanks
| 1
|
288,708
| 8,850,587,584
|
IssuesEvent
|
2019-01-08 13:41:30
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
jobs.jobvite.com - design is broken
|
browser-firefox priority-normal
|
<!-- @browser: Firefox 65.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:65.0) Gecko/20100101 Firefox/65.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed
**Browser / Version**: Firefox 65.0
**Operating System**: Linux
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Site will not submit application
**Steps to Reproduce**:
Last button does not function.
[](https://webcompat.com/uploads/2018/12/eddec568-ea29-42c3-bcad-aca71a843db1.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181207174328</li><li>tracking content blocked: true (strict)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "unreachable code after return statement" {file: "http://app.jvicdn.com/__assets__/lib/moment/moment.min.js?v=181213135315" line: 6 column: 12432 source: "e._d.setTime(+e._d+f),a.updateOffset(e,!1),e):Aa(b).local();return c._isUTC?Aa(b).zone(c._offset||0):Aa(b).local()}funct"}]', u'[JavaScript Warning: "The resource at http://www.google-analytics.com/ga.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://www.google-analytics.com/ga.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 35}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.linkedin.com/in.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 346}]', u'[JavaScript Error: "The stylesheet http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed was not loaded because its MIME type, text/html, is not text/css." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "The resource at http://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.twitter.com/widgets.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[JavaScript Warning: "The resource at http://connect.facebook.net/en_US/sdk.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://connect.facebook.net/en_US/sdk.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[console.log(Replacing social widget for LinkedIn) moz-extension://921d6c32-4cd8-46d4-93b8-c67e37371904/js/contentscripts/socialwidgets.js:306:5]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "moz-extension://381126cf-dbe6-450d-8097-a68a2e0a1032/content_scripts/handler.js" line: 73}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
jobs.jobvite.com - design is broken - <!-- @browser: Firefox 65.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:65.0) Gecko/20100101 Firefox/65.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed
**Browser / Version**: Firefox 65.0
**Operating System**: Linux
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Site will not submit application
**Steps to Reproduce**:
Last button does not function.
[](https://webcompat.com/uploads/2018/12/eddec568-ea29-42c3-bcad-aca71a843db1.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181207174328</li><li>tracking content blocked: true (strict)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "unreachable code after return statement" {file: "http://app.jvicdn.com/__assets__/lib/moment/moment.min.js?v=181213135315" line: 6 column: 12432 source: "e._d.setTime(+e._d+f),a.updateOffset(e,!1),e):Aa(b).local();return c._isUTC?Aa(b).zone(c._offset||0):Aa(b).local()}funct"}]', u'[JavaScript Warning: "The resource at http://www.google-analytics.com/ga.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://www.google-analytics.com/ga.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 35}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.linkedin.com/in.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 346}]', u'[JavaScript Error: "The stylesheet http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed was not loaded because its MIME type, text/html, is not text/css." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "The resource at http://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://platform.twitter.com/widgets.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[JavaScript Warning: "The resource at http://connect.facebook.net/en_US/sdk.js was blocked because content blocking is enabled." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source http://connect.facebook.net/en_US/sdk.js." {file: "http://jobs.jobvite.com/careers/jbs/job/oTp08fwa?__jvst=Job%20Board&__jvsd=Indeed" line: 1}]', u'[console.log(Replacing social widget for LinkedIn) moz-extension://921d6c32-4cd8-46d4-93b8-c67e37371904/js/contentscripts/socialwidgets.js:306:5]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "moz-extension://381126cf-dbe6-450d-8097-a68a2e0a1032/content_scripts/handler.js" line: 73}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
jobs jobvite com design is broken url browser version firefox operating system linux tested another browser yes problem type design is broken description site will not submit application steps to reproduce last button does not function browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked true strict gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel aurora console messages u u u u u u u u u u from with ❤️
| 0
|
674,133
| 23,040,439,666
|
IssuesEvent
|
2022-07-23 04:00:40
|
space-wizards/space-station-14
|
https://api.github.com/repos/space-wizards/space-station-14
|
closed
|
Shuttle/vehicle movement structure seems dodgy
|
Priority: 3-Not Required Needs Discussion Difficulty: 3-Hard
|
## Description
This issue is very much an abstract code-structure issue right now, but I think it'll be a problem later.
I was told to open an issue on this.
Right now, there is no actual concept of movement states (i.e. no interface for such).
However, there are three movement modes regardless. They are laid out in an order which is not respective of priority in code that otherwise seems like it's supposed to use this information to achieve prioritization (server `MoverController`):
+ Mob Movement - Always applied if `IMobMoverComponent` is available. Adds to `_excludedMobs`, does not read. Standard movement. Cares about CanMove (but this is 'interior', not relevant for mode prioritization)
+ Shuttle Movement - Applied if `PilotComponent` is available *and* it is connected to a console. Adds to `_excludedMobs`, does not read. Shuttle control.
+ Kinematic Movement - Always applied if `IMoverComponent` (required for any movement processing whatsoever) is available. Reads and ignores those entities in `_excludedMobs`. Ghost movement - ignores CanMove, etc.
This points to some attempt to use `_excludedMobs` to use prioritization to keep the states consistent, but this wasn't actually made use of properly, so instead `PilotComponent` is also a marker that enacts a movement block if it's connected to a console.
## Specific concerns
The specific concerns I have with this approach are that adding states to this will inevitably require those states trigger movement blocks, which will prevent *other* movement blocks from being tested for even when they really should be.
(In particular, I was looking into how this might have an effect on the development of ~~clown cars~~ non-shuttle vehicles. The answer appears to be that to implement those, one would need to either handle it via buckling (likely not an option for various reasons under normal circumstances) or implement some sort of mirror system to piloting for the driver of a vehicle, doing a movement block, etc.
The worry I have is that doing such a thing would interact badly with if someone driving a vehicle then went on to attempt to pilot a shuttle while driving a vehicle. In this case, both movement states would then be in effect simultaneously.)
## Citations
Relevant code (as of master at time of writing) with their roles in shuttle control is here:
+ Ignores shuttle control but gets disabled by movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Client/Physics/Controllers/MoverController.cs
+ Contains the action blocker call which causes movement attempt events: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Movement/SharedMoverController.cs
+ The actual server-side authoritative movement logic, including the priority-disable of kinematic movement: https://github.com/space-wizards/space-station-14/blob/master/Content.Server/Physics/Controllers/MoverController.cs
+ Shuttle movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Shuttles/SharedShuttleConsoleSystem.cs
|
1.0
|
Shuttle/vehicle movement structure seems dodgy - ## Description
This issue is very much an abstract code-structure issue right now, but I think it'll be a problem later.
I was told to open an issue on this.
Right now, there is no actual concept of movement states (i.e. no interface for such).
However, there are three movement modes regardless. They are laid out in an order which is not respective of priority in code that otherwise seems like it's supposed to use this information to achieve prioritization (server `MoverController`):
+ Mob Movement - Always applied if `IMobMoverComponent` is available. Adds to `_excludedMobs`, does not read. Standard movement. Cares about CanMove (but this is 'interior', not relevant for mode prioritization)
+ Shuttle Movement - Applied if `PilotComponent` is available *and* it is connected to a console. Adds to `_excludedMobs`, does not read. Shuttle control.
+ Kinematic Movement - Always applied if `IMoverComponent` (required for any movement processing whatsoever) is available. Reads and ignores those entities in `_excludedMobs`. Ghost movement - ignores CanMove, etc.
This points to some attempt to use `_excludedMobs` to use prioritization to keep the states consistent, but this wasn't actually made use of properly, so instead `PilotComponent` is also a marker that enacts a movement block if it's connected to a console.
## Specific concerns
The specific concerns I have with this approach are that adding states to this will inevitably require those states trigger movement blocks, which will prevent *other* movement blocks from being tested for even when they really should be.
(In particular, I was looking into how this might have an effect on the development of ~~clown cars~~ non-shuttle vehicles. The answer appears to be that to implement those, one would need to either handle it via buckling (likely not an option for various reasons under normal circumstances) or implement some sort of mirror system to piloting for the driver of a vehicle, doing a movement block, etc.
The worry I have is that doing such a thing would interact badly with if someone driving a vehicle then went on to attempt to pilot a shuttle while driving a vehicle. In this case, both movement states would then be in effect simultaneously.)
## Citations
Relevant code (as of master at time of writing) with their roles in shuttle control is here:
+ Ignores shuttle control but gets disabled by movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Client/Physics/Controllers/MoverController.cs
+ Contains the action blocker call which causes movement attempt events: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Movement/SharedMoverController.cs
+ The actual server-side authoritative movement logic, including the priority-disable of kinematic movement: https://github.com/space-wizards/space-station-14/blob/master/Content.Server/Physics/Controllers/MoverController.cs
+ Shuttle movement blocker: https://github.com/space-wizards/space-station-14/blob/master/Content.Shared/Shuttles/SharedShuttleConsoleSystem.cs
|
non_process
|
shuttle vehicle movement structure seems dodgy description this issue is very much an abstract code structure issue right now but i think it ll be a problem later i was told to open an issue on this right now there is no actual concept of movement states i e no interface for such however there are three movement modes regardless they are laid out in an order which is not respective of priority in code that otherwise seems like it s supposed to use this information to achieve prioritization server movercontroller mob movement always applied if imobmovercomponent is available adds to excludedmobs does not read standard movement cares about canmove but this is interior not relevant for mode prioritization shuttle movement applied if pilotcomponent is available and it is connected to a console adds to excludedmobs does not read shuttle control kinematic movement always applied if imovercomponent required for any movement processing whatsoever is available reads and ignores those entities in excludedmobs ghost movement ignores canmove etc this points to some attempt to use excludedmobs to use prioritization to keep the states consistent but this wasn t actually made use of properly so instead pilotcomponent is also a marker that enacts a movement block if it s connected to a console specific concerns the specific concerns i have with this approach are that adding states to this will inevitably require those states trigger movement blocks which will prevent other movement blocks from being tested for even when they really should be in particular i was looking into how this might have an effect on the development of clown cars non shuttle vehicles the answer appears to be that to implement those one would need to either handle it via buckling likely not an option for various reasons under normal circumstances or implement some sort of mirror system to piloting for the driver of a vehicle doing a movement block etc the worry i have is that doing such a thing would interact badly with if someone driving a vehicle then went on to attempt to pilot a shuttle while driving a vehicle in this case both movement states would then be in effect simultaneously citations relevant code as of master at time of writing with their roles in shuttle control is here ignores shuttle control but gets disabled by movement blocker contains the action blocker call which causes movement attempt events the actual server side authoritative movement logic including the priority disable of kinematic movement shuttle movement blocker
| 0
|
13,506
| 16,045,207,919
|
IssuesEvent
|
2021-04-22 12:55:48
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Prisma Format: Fill in native type attributes on foreign key relations
|
kind/bug process/candidate team/client topic: mongodb
|
## Problem
This problem popped up while working with Mongo. Given the following:
```prisma
datasource db {
provider = "mongodb"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
previewFeatures = ["mongodb"]
}
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
}
```
All good so far, when we go to link these two models with User posts:
```prisma
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
posts Post[]
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
}
```
It auto-formats to:
```prisma
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
posts Post[]
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
User User? @relation(fields: [userId], references: [id])
userId String?
}
```
There's a red line under `User User?` with the following:
```
Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`.
```
You can also see this when you run `npx prisma format`:
```
Environment variables loaded from .env
Prisma schema loaded from prisma/schema.prisma
Error: Schema parsing
error: Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`.
--> schema.prisma:21
|
21 | User User? @relation(fields: [userId], references: [id])
22 | userId String?
|
Validation Error Count: 1
```
This is slightly opaque, but the solution is to add `@db.ObjectId` to `userId` like so:
```prisma
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
posts Post[]
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
User User? @relation(fields: [userId], references: [id])
userId String? @db.ObjectId
}
```
|
1.0
|
Prisma Format: Fill in native type attributes on foreign key relations - ## Problem
This problem popped up while working with Mongo. Given the following:
```prisma
datasource db {
provider = "mongodb"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
previewFeatures = ["mongodb"]
}
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
}
```
All good so far, when we go to link these two models with User posts:
```prisma
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
posts Post[]
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
}
```
It auto-formats to:
```prisma
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
posts Post[]
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
User User? @relation(fields: [userId], references: [id])
userId String?
}
```
There's a red line under `User User?` with the following:
```
Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`.
```
You can also see this when you run `npx prisma format`:
```
Environment variables loaded from .env
Prisma schema loaded from prisma/schema.prisma
Error: Schema parsing
error: Error parsing attribute "@relation": The type of the field `userId` in the model `Post` is not matching the type of the referenced field `id` in model `User`.
--> schema.prisma:21
|
21 | User User? @relation(fields: [userId], references: [id])
22 | userId String?
|
Validation Error Count: 1
```
This is slightly opaque, but the solution is to add `@db.ObjectId` to `userId` like so:
```prisma
model User {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
name String
posts Post[]
}
model Post {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
title String
User User? @relation(fields: [userId], references: [id])
userId String? @db.ObjectId
}
```
|
process
|
prisma format fill in native type attributes on foreign key relations problem this problem popped up while working with mongo given the following prisma datasource db provider mongodb url env database url generator client provider prisma client js previewfeatures model user id string id default dbgenerated map id db objectid name string model post id string id default dbgenerated map id db objectid title string all good so far when we go to link these two models with user posts prisma model user id string id default dbgenerated map id db objectid name string posts post model post id string id default dbgenerated map id db objectid title string it auto formats to prisma model user id string id default dbgenerated map id db objectid name string posts post model post id string id default dbgenerated map id db objectid title string user user relation fields references userid string there s a red line under user user with the following error parsing attribute relation the type of the field userid in the model post is not matching the type of the referenced field id in model user you can also see this when you run npx prisma format environment variables loaded from env prisma schema loaded from prisma schema prisma error schema parsing error error parsing attribute relation the type of the field userid in the model post is not matching the type of the referenced field id in model user schema prisma user user relation fields references userid string validation error count this is slightly opaque but the solution is to add db objectid to userid like so prisma model user id string id default dbgenerated map id db objectid name string posts post model post id string id default dbgenerated map id db objectid title string user user relation fields references userid string db objectid
| 1
|
212,504
| 16,487,440,007
|
IssuesEvent
|
2021-05-24 20:16:25
|
anitab-org/stem-diverse-tv
|
https://api.github.com/repos/anitab-org/stem-diverse-tv
|
closed
|
Bug : Incorrect link to bug_report.md in contributing_guidelines.md
|
Category: Documentation/Training First Timers Only Type: Bug
|
### Describe the bug
The links to bug report is incorrect in the Contributing Guidelines file you can see this by also scrolling to the [Contributing section](https://github.com/anitab-org/stem-diverse-tv#contributing) in readme
### To Reproduce
Steps to reproduce the behavior:
1. Go to 'https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/contributing_guidelines.md'
2. Scroll down to Contribution guidelines section
3. Click on [bug_report.md](https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/ISSUE_TEMPLATE/bug_report.md)
4. See error
### Expected behavior
Links to the file should be corrected
bug report: https://github.com/anitab-org/stem-diverse-tv/blob/master/.github/ISSUE_TEMPLATE/bug_report.md
### Desktop (please complete the following information):
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
### Smartphone (please complete the following information):
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
### Additional context
Add any other context about the problem here.
|
1.0
|
Bug : Incorrect link to bug_report.md in contributing_guidelines.md - ### Describe the bug
The links to bug report is incorrect in the Contributing Guidelines file you can see this by also scrolling to the [Contributing section](https://github.com/anitab-org/stem-diverse-tv#contributing) in readme
### To Reproduce
Steps to reproduce the behavior:
1. Go to 'https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/contributing_guidelines.md'
2. Scroll down to Contribution guidelines section
3. Click on [bug_report.md](https://github.com/anitab-org/stem-diverse-tv/blob/master/docs/ISSUE_TEMPLATE/bug_report.md)
4. See error
### Expected behavior
Links to the file should be corrected
bug report: https://github.com/anitab-org/stem-diverse-tv/blob/master/.github/ISSUE_TEMPLATE/bug_report.md
### Desktop (please complete the following information):
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
### Smartphone (please complete the following information):
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
### Additional context
Add any other context about the problem here.
|
non_process
|
bug incorrect link to bug report md in contributing guidelines md describe the bug the links to bug report is incorrect in the contributing guidelines file you can see this by also scrolling to the in readme to reproduce steps to reproduce the behavior go to scroll down to contribution guidelines section click on see error expected behavior links to the file should be corrected bug report desktop please complete the following information os browser version smartphone please complete the following information device os browser version additional context add any other context about the problem here
| 0
|
344,704
| 24,824,023,485
|
IssuesEvent
|
2022-10-25 18:56:29
|
pnp/powershell
|
https://api.github.com/repos/pnp/powershell
|
closed
|
Documentation: Using pnp/powershell Docker Image
|
documentation
|
**Is your feature request related to a problem? Please describe.**
Starting using pnp/powershell Docker Image might be tricky for beginners
**Describe the solution you'd like**
An article, similar to https://pnp.github.io/powershell/articles/azurecloudshell.html, that presents different options and different examples of using the module in Docker.
**Describe alternatives you've considered**
Alternative could be experimenting and learning individually
**Additional context**
|
1.0
|
Documentation: Using pnp/powershell Docker Image - **Is your feature request related to a problem? Please describe.**
Starting using pnp/powershell Docker Image might be tricky for beginners
**Describe the solution you'd like**
An article, similar to https://pnp.github.io/powershell/articles/azurecloudshell.html, that presents different options and different examples of using the module in Docker.
**Describe alternatives you've considered**
Alternative could be experimenting and learning individually
**Additional context**
|
non_process
|
documentation using pnp powershell docker image is your feature request related to a problem please describe starting using pnp powershell docker image might be tricky for beginners describe the solution you d like an article similar to that presents different options and different examples of using the module in docker describe alternatives you ve considered alternative could be experimenting and learning individually additional context
| 0
|
14,913
| 18,296,692,751
|
IssuesEvent
|
2021-10-05 21:11:06
|
pavel-one/Logger
|
https://api.github.com/repos/pavel-one/Logger
|
opened
|
Регистрация и вход ч/з соцсети
|
Process
|
Сделать регистрацию через следующие соцсети:
1. Github
2. Google
|
1.0
|
Регистрация и вход ч/з соцсети - Сделать регистрацию через следующие соцсети:
1. Github
2. Google
|
process
|
регистрация и вход ч з соцсети сделать регистрацию через следующие соцсети github google
| 1
|
9,642
| 12,603,535,545
|
IssuesEvent
|
2020-06-11 13:38:21
|
HackYourFutureBelgium/class-9-10
|
https://api.github.com/repos/HackYourFutureBelgium/class-9-10
|
opened
|
Your Name: module, week
|
class-10 process-week wednesday-check-in
|
# Wednesday Check-In
__Debugging, Week 1__
## Progress
I learned about JavaScript concepts.
## Blocked
So far the process is clear.
## Next Steps
Learn more about JavaScript since it requires more practicing to understand it very well.
## Tip(s) of the week
I watched videos on YouTube and they were very useful, but they were in Arabic 🤓
|
1.0
|
Your Name: module, week - # Wednesday Check-In
__Debugging, Week 1__
## Progress
I learned about JavaScript concepts.
## Blocked
So far the process is clear.
## Next Steps
Learn more about JavaScript since it requires more practicing to understand it very well.
## Tip(s) of the week
I watched videos on YouTube and they were very useful, but they were in Arabic 🤓
|
process
|
your name module week wednesday check in debugging week progress i learned about javascript concepts blocked so far the process is clear next steps learn more about javascript since it requires more practicing to understand it very well tip s of the week i watched videos on youtube and they were very useful but they were in arabic 🤓
| 1
|
17,367
| 23,191,007,326
|
IssuesEvent
|
2022-08-01 12:40:05
|
apache/arrow-rs
|
https://api.github.com/repos/apache/arrow-rs
|
opened
|
Release `object_store` `0.4.0`
|
development-process
|
As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180
* Planned Release Candidate: 2022-08-08
* Planned Release and Publish to crates.io: 2022-08-11
Items:
- [ ] Update changelog and readme:
- [ ] Create release scripts
- [ ] Create release candidate
- [ ] Release candidate approved
- [ ] Release to crates.io
- [ ] Create release scripts
|
1.0
|
Release `object_store` `0.4.0` - As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180
* Planned Release Candidate: 2022-08-08
* Planned Release and Publish to crates.io: 2022-08-11
Items:
- [ ] Update changelog and readme:
- [ ] Create release scripts
- [ ] Create release candidate
- [ ] Release candidate approved
- [ ] Release to crates.io
- [ ] Create release scripts
|
process
|
release object store as discussed we will release a new version of object store from the arrow rs repo under asf process planned release candidate planned release and publish to crates io items update changelog and readme create release scripts create release candidate release candidate approved release to crates io create release scripts
| 1
|
18,042
| 24,052,944,827
|
IssuesEvent
|
2022-09-16 14:19:03
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
closed
|
Add IERC transferFrom to htsPrecompile approval acceptance tests
|
enhancement limechain P1 process
|
### Problem
Currently the `htsPrecompile.spec.ts` has tests for approval flow.
However, it doesn't follow up with actual transfer tests.
### Solution
Add an additional step to each `approval` test to test the `transferFrom()` call.
Ensure both `IERC20` and `IERC721` get coverage
- `IERC20(token).transferFrom(sender, recipient, amount);`
- `IERC721(token).transferFrom(from, to, tokenId);`
### Alternatives
_No response_
|
1.0
|
Add IERC transferFrom to htsPrecompile approval acceptance tests - ### Problem
Currently the `htsPrecompile.spec.ts` has tests for approval flow.
However, it doesn't follow up with actual transfer tests.
### Solution
Add an additional step to each `approval` test to test the `transferFrom()` call.
Ensure both `IERC20` and `IERC721` get coverage
- `IERC20(token).transferFrom(sender, recipient, amount);`
- `IERC721(token).transferFrom(from, to, tokenId);`
### Alternatives
_No response_
|
process
|
add ierc transferfrom to htsprecompile approval acceptance tests problem currently the htsprecompile spec ts has tests for approval flow however it doesn t follow up with actual transfer tests solution add an additional step to each approval test to test the transferfrom call ensure both and get coverage token transferfrom sender recipient amount token transferfrom from to tokenid alternatives no response
| 1
|
11,884
| 14,680,446,267
|
IssuesEvent
|
2020-12-31 10:05:46
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] [Dev] Studies created from SB having maximum Research Sponsor text are not reflected in Sites/Studies tab in PM
|
Bug P1 Participant manager Process: Dev Process: Tested QA
|
Steps:
1. Create a study from SB by giving maximum Research Sponsor text in basic information
2. Launch the study
3. Navigate to PM --> Sites tab
4. Observe the list
A/R: Some of the studies created from SB are not reflected in Sites/Studies tab in PM
E/R: All studies created from SB should be reflected
Study details which is not reflected in PM
Study Name: Clinical Trial of the Moventis PNS for Knee Pain
Study ID: KneePain
App ID: BTCDEV001
Instance: Dev
|
2.0
|
[PM] [Dev] Studies created from SB having maximum Research Sponsor text are not reflected in Sites/Studies tab in PM - Steps:
1. Create a study from SB by giving maximum Research Sponsor text in basic information
2. Launch the study
3. Navigate to PM --> Sites tab
4. Observe the list
A/R: Some of the studies created from SB are not reflected in Sites/Studies tab in PM
E/R: All studies created from SB should be reflected
Study details which is not reflected in PM
Study Name: Clinical Trial of the Moventis PNS for Knee Pain
Study ID: KneePain
App ID: BTCDEV001
Instance: Dev
|
process
|
studies created from sb having maximum research sponsor text are not reflected in sites studies tab in pm steps create a study from sb by giving maximum research sponsor text in basic information launch the study navigate to pm sites tab observe the list a r some of the studies created from sb are not reflected in sites studies tab in pm e r all studies created from sb should be reflected study details which is not reflected in pm study name clinical trial of the moventis pns for knee pain study id kneepain app id instance dev
| 1
|
139,368
| 18,850,349,884
|
IssuesEvent
|
2021-11-11 19:59:18
|
snowdensb/sonar-xanitizer
|
https://api.github.com/repos/snowdensb/sonar-xanitizer
|
opened
|
CVE-2020-24616 (High) detected in jackson-databind-2.6.3.jar
|
security vulnerability
|
## CVE-2020-24616 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /src/test/resources/webgoat/WEB-INF/lib/jackson-databind-2.6.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/sonar-xanitizer/commit/e2144e84b1fdbf18c01c24e6ab9ade7b45b25283">e2144e84b1fdbf18c01c24e6ab9ade7b45b25283</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).
<p>Publish Date: 2020-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-25</p>
<p>Fix Resolution: 2.9.10.6</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.3","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-24616 (High) detected in jackson-databind-2.6.3.jar - ## CVE-2020-24616 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /src/test/resources/webgoat/WEB-INF/lib/jackson-databind-2.6.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/sonar-xanitizer/commit/e2144e84b1fdbf18c01c24e6ab9ade7b45b25283">e2144e84b1fdbf18c01c24e6ab9ade7b45b25283</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).
<p>Publish Date: 2020-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-25</p>
<p>Fix Resolution: 2.9.10.6</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.3","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library src test resources webgoat web inf lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp vulnerabilityurl
| 0
|
3,728
| 6,733,142,338
|
IssuesEvent
|
2017-10-18 13:58:37
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
closed
|
Contract Preparation Detail Wizard
|
form process workflow
|
Two Steps Wizard Wireframe

Dynamically load fields that are applicable for the chosen activity type.
**After submitted, only update in the collection**
|
1.0
|
Contract Preparation Detail Wizard - Two Steps Wizard Wireframe

Dynamically load fields that are applicable for the chosen activity type.
**After submitted, only update in the collection**
|
process
|
contract preparation detail wizard two steps wizard wireframe dynamically load fields that are applicable for the chosen activity type after submitted only update in the collection
| 1
|
614,226
| 19,161,598,265
|
IssuesEvent
|
2021-12-03 01:15:45
|
returntocorp/semgrep
|
https://api.github.com/repos/returntocorp/semgrep
|
closed
|
Opposite conditional matching behavior in PHP and Java
|
priority:low lang:java lang:php feature:matching stale
|
In Java matching conditionals with
```
if (...) { ... }
```
matches an if-block as well as an if-else-block. In PHP only the if-block is matched.
However, changing the pattern to
```
if (...) { ... } else { ... }
```
will have the opposite behavior in either languages.
Links:
https://semgrep.dev/s/bashprince:java
https://semgrep.dev/s/bashprince:php
Not sure which one is the correct behavior, but I would think the Java version makes more sense in this case since it matches the whole block and the further specification with else makes the match more exact.
- [ ] P2: regular bug that should get fixed
Tested on semgrep.dev
|
1.0
|
Opposite conditional matching behavior in PHP and Java - In Java matching conditionals with
```
if (...) { ... }
```
matches an if-block as well as an if-else-block. In PHP only the if-block is matched.
However, changing the pattern to
```
if (...) { ... } else { ... }
```
will have the opposite behavior in either languages.
Links:
https://semgrep.dev/s/bashprince:java
https://semgrep.dev/s/bashprince:php
Not sure which one is the correct behavior, but I would think the Java version makes more sense in this case since it matches the whole block and the further specification with else makes the match more exact.
- [ ] P2: regular bug that should get fixed
Tested on semgrep.dev
|
non_process
|
opposite conditional matching behavior in php and java in java matching conditionals with if matches an if block as well as an if else block in php only the if block is matched however changing the pattern to if else will have the opposite behavior in either languages links not sure which one is the correct behavior but i would think the java version makes more sense in this case since it matches the whole block and the further specification with else makes the match more exact regular bug that should get fixed tested on semgrep dev
| 0
|
121,701
| 16,016,616,584
|
IssuesEvent
|
2021-04-20 16:48:04
|
microsoft/BotFramework-Composer
|
https://api.github.com/repos/microsoft/BotFramework-Composer
|
closed
|
Restart selected bot with single click
|
Bot Services UX Design customer-reported feature-request
|
## Is your feature request related to a problem? Please describe.
#7049
If multiple bot is running , there is no easy solution to restart the particular bot , current approach we have restart all the bots or stop and start selected bot.

## Describe the solution you'd like
If user has selected overall project

If user selected particular bot then restart only particular bot instead of all


|
1.0
|
Restart selected bot with single click - ## Is your feature request related to a problem? Please describe.
#7049
If multiple bot is running , there is no easy solution to restart the particular bot , current approach we have restart all the bots or stop and start selected bot.

## Describe the solution you'd like
If user has selected overall project

If user selected particular bot then restart only particular bot instead of all


|
non_process
|
restart selected bot with single click is your feature request related to a problem please describe if multiple bot is running there is no easy solution to restart the particular bot current approach we have restart all the bots or stop and start selected bot describe the solution you d like if user has selected overall project if user selected particular bot then restart only particular bot instead of all
| 0
|
434,577
| 30,456,278,918
|
IssuesEvent
|
2023-07-16 23:12:32
|
CarsOk/droguerias_darcy
|
https://api.github.com/repos/CarsOk/droguerias_darcy
|
closed
|
Diagrama entidad relacion
|
documentation
|
## Yo como
Aprendiz SENA
## Quiero
Subir el diagrama entidad relacion del proyecto
## Para
Seguimiento inicial
|
1.0
|
Diagrama entidad relacion - ## Yo como
Aprendiz SENA
## Quiero
Subir el diagrama entidad relacion del proyecto
## Para
Seguimiento inicial
|
non_process
|
diagrama entidad relacion yo como aprendiz sena quiero subir el diagrama entidad relacion del proyecto para seguimiento inicial
| 0
|
87,711
| 25,192,032,325
|
IssuesEvent
|
2022-11-12 03:28:19
|
tensorflow/tfjs
|
https://api.github.com/repos/tensorflow/tfjs
|
closed
|
Compile error with @tensorflow/tfjs-backend-webgl for TS target > ES5
|
type:build/install
|
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: -
- TensorFlow.js installed from (npm or script link): npm
- TensorFlow.js version: 2.7.0
- CUDA/cuDNN version: -
**Describe the problem**
When trying to build a typescript code that depends on `@tensorflow/tfjs-backend-webgl`, if the `compilerOptions.target` is above ES5 (i.e "ES2015","ES2016","ES2017","ES2018","ES2019","ES2020" or "ESNext"), you get the following compilation errors:
```
node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextBase'.
Types of property 'clearBufferfv' are incompatible.
Type '(buffer: number, drawbuffer: number, values: ArrayLike<number> | Float32Array, srcOffset?: number) => void' is not assignable to type '{ (buffer: number, drawbuffer: number, values: Float32List, srcOffset?: number): void; (buffer: number, drawbuffer: number, values: Iterable<number>, srcOffset?: number): void; }'.
Types of parameters 'values' and 'values' are incompatible.
Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'.
Type 'Iterable<number>' is missing the following properties from type 'Float32Array': BYTES_PER_ELEMENT, buffer, byteLength, byteOffset, and 25 more.
16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase {
~~~~~~~~~~~~~~~~~~~~~~
node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextOverloads'.
Types of property 'uniform1fv' are incompatible.
Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, data: Float32List, srcOffset?: number, srcLength?: number): void; (location: WebGLUniformLocation, data: Iterable<...>, srcOffset?: number, srcLength?: number): void; }'.
Types of parameters 'data' and 'data' are incompatible.
Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'.
Type 'Iterable<number>' is not assignable to type 'Float32Array'.
16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase {
~~~~~~~~~~~~~~~~~~~~~~
node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGLRenderingContext'.
Types of property 'uniform1fv' are incompatible.
Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, v: Float32List): void; (location: WebGLUniformLocation, v: Iterable<number>): void; }'.
Types of parameters 'data' and 'v' are incompatible.
Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'.
Type 'Iterable<number>' is not assignable to type 'Float32Array'.
16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase {
~~~~~~~~~~~~~~~~~~~~~~
Found 3 errors.
```
**Provide the exact sequence of commands / steps that you executed before running into the problem**
The error is easily reconstructed with the following:
1. package.json:
```
{
"name": "tfjs-ts-test",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"build": "tsc"
},
"author": "",
"license": "ISC",
"dependencies": {
"@tensorflow/tfjs-backend-webgl": "^2.7.0",
"@tensorflow/tfjs-core": "^2.7.0",
"typescript": "^4.0.5"
}
}
```
2. tsconfig.json:
```
{
"compilerOptions": {
"target": "ES2015"
}
}
```
3. An empty index.ts
4. npm install
5. npm run build
**Any other info / logs**
This partially relates to https://github.com/tensorflow/tfjs/issues/2007 but not the same issue
|
1.0
|
Compile error with @tensorflow/tfjs-backend-webgl for TS target > ES5 - **System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 10
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: -
- TensorFlow.js installed from (npm or script link): npm
- TensorFlow.js version: 2.7.0
- CUDA/cuDNN version: -
**Describe the problem**
When trying to build a typescript code that depends on `@tensorflow/tfjs-backend-webgl`, if the `compilerOptions.target` is above ES5 (i.e "ES2015","ES2016","ES2017","ES2018","ES2019","ES2020" or "ESNext"), you get the following compilation errors:
```
node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextBase'.
Types of property 'clearBufferfv' are incompatible.
Type '(buffer: number, drawbuffer: number, values: ArrayLike<number> | Float32Array, srcOffset?: number) => void' is not assignable to type '{ (buffer: number, drawbuffer: number, values: Float32List, srcOffset?: number): void; (buffer: number, drawbuffer: number, values: Iterable<number>, srcOffset?: number): void; }'.
Types of parameters 'values' and 'values' are incompatible.
Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'.
Type 'Iterable<number>' is missing the following properties from type 'Float32Array': BYTES_PER_ELEMENT, buffer, byteLength, byteOffset, and 25 more.
16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase {
~~~~~~~~~~~~~~~~~~~~~~
node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGL2RenderingContextOverloads'.
Types of property 'uniform1fv' are incompatible.
Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, data: Float32List, srcOffset?: number, srcLength?: number): void; (location: WebGLUniformLocation, data: Iterable<...>, srcOffset?: number, srcLength?: number): void; }'.
Types of parameters 'data' and 'data' are incompatible.
Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'.
Type 'Iterable<number>' is not assignable to type 'Float32Array'.
16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase {
~~~~~~~~~~~~~~~~~~~~~~
node_modules/typescript/lib/lib.dom.d.ts:16532:11 - error TS2430: Interface 'WebGL2RenderingContext' incorrectly extends interface 'WebGLRenderingContext'.
Types of property 'uniform1fv' are incompatible.
Type '(location: WebGLUniformLocation, data: ArrayLike<number> | Float32Array, srcOffset?: number, srcLength?: number) => void' is not assignable to type '{ (location: WebGLUniformLocation, v: Float32List): void; (location: WebGLUniformLocation, v: Iterable<number>): void; }'.
Types of parameters 'data' and 'v' are incompatible.
Type 'Iterable<number>' is not assignable to type 'ArrayLike<number> | Float32Array'.
Type 'Iterable<number>' is not assignable to type 'Float32Array'.
16532 interface WebGL2RenderingContext extends WebGL2RenderingContextBase, WebGL2RenderingContextOverloads, WebGLRenderingContextBase {
~~~~~~~~~~~~~~~~~~~~~~
Found 3 errors.
```
**Provide the exact sequence of commands / steps that you executed before running into the problem**
The error is easily reconstructed with the following:
1. package.json:
```
{
"name": "tfjs-ts-test",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"build": "tsc"
},
"author": "",
"license": "ISC",
"dependencies": {
"@tensorflow/tfjs-backend-webgl": "^2.7.0",
"@tensorflow/tfjs-core": "^2.7.0",
"typescript": "^4.0.5"
}
}
```
2. tsconfig.json:
```
{
"compilerOptions": {
"target": "ES2015"
}
}
```
3. An empty index.ts
4. npm install
5. npm run build
**Any other info / logs**
This partially relates to https://github.com/tensorflow/tfjs/issues/2007 but not the same issue
|
non_process
|
compile error with tensorflow tfjs backend webgl for ts target system information os platform and distribution e g linux ubuntu windows mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow js installed from npm or script link npm tensorflow js version cuda cudnn version describe the problem when trying to build a typescript code that depends on tensorflow tfjs backend webgl if the compileroptions target is above i e or esnext you get the following compilation errors node modules typescript lib lib dom d ts error interface incorrectly extends interface types of property clearbufferfv are incompatible type buffer number drawbuffer number values arraylike srcoffset number void is not assignable to type buffer number drawbuffer number values srcoffset number void buffer number drawbuffer number values iterable srcoffset number void types of parameters values and values are incompatible type iterable is not assignable to type arraylike type iterable is missing the following properties from type bytes per element buffer bytelength byteoffset and more interface extends webglrenderingcontextbase node modules typescript lib lib dom d ts error interface incorrectly extends interface types of property are incompatible type location webgluniformlocation data arraylike srcoffset number srclength number void is not assignable to type location webgluniformlocation data srcoffset number srclength number void location webgluniformlocation data iterable srcoffset number srclength number void types of parameters data and data are incompatible type iterable is not assignable to type arraylike type iterable is not assignable to type interface extends webglrenderingcontextbase node modules typescript lib lib dom d ts error interface incorrectly extends interface webglrenderingcontext types of property are incompatible type location webgluniformlocation data arraylike srcoffset number srclength number void is not assignable to type location webgluniformlocation v void location webgluniformlocation v iterable void types of parameters data and v are incompatible type iterable is not assignable to type arraylike type iterable is not assignable to type interface extends webglrenderingcontextbase found errors provide the exact sequence of commands steps that you executed before running into the problem the error is easily reconstructed with the following package json name tfjs ts test version description main index js scripts build tsc author license isc dependencies tensorflow tfjs backend webgl tensorflow tfjs core typescript tsconfig json compileroptions target an empty index ts npm install npm run build any other info logs this partially relates to but not the same issue
| 0
|
258,224
| 27,563,872,095
|
IssuesEvent
|
2023-03-08 01:12:29
|
billmcchesney1/t-vault
|
https://api.github.com/repos/billmcchesney1/t-vault
|
opened
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz
|
security vulnerability
|
## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: /tvaultui/package.json</p>
<p>Path to vulnerable library: /tvaultui/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-3.1.0.tgz (Root Library)
- node-sass-4.14.1.tgz
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution (hosted-git-info): 2.8.9</p>
<p>Direct dependency fix Resolution (gulp-sass): 3.2.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
True
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: /tvaultui/package.json</p>
<p>Path to vulnerable library: /tvaultui/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-3.1.0.tgz (Root Library)
- node-sass-4.14.1.tgz
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution (hosted-git-info): 2.8.9</p>
<p>Direct dependency fix Resolution (gulp-sass): 3.2.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
non_process
|
cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file tvaultui package json path to vulnerable library tvaultui node modules hosted git info package json dependency hierarchy gulp sass tgz root library node sass tgz meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in base branch dev vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info direct dependency fix resolution gulp sass check this box to open an automated fix pr
| 0
|
40,869
| 10,586,892,721
|
IssuesEvent
|
2019-10-08 20:45:33
|
ged/ruby-pg
|
https://api.github.com/repos/ged/ruby-pg
|
closed
|
gem install pg on windows 7 fails
|
0.10.1 Build System bug major
|
**[Original report](https://bitbucket.org/ged/ruby-pg/issue/64) by Anonymous.**
----------------------------------------
When I run "bundle install", I get this error:
checking for pg_config... yes
Using config values from c:\PostgreSQL\9.0\bin/pg_config.exe
checking for libpq-fe.h... yes
checking for libpq/libpq-fs.h... yes
checking for PQconnectdb() in -lpq... no
checking for PQconnectdb() in -llibpq... no
checking for PQconnectdb() in -lms/libpq... no
Can't find the PostgreSQL client library (libpq)
*** extconf.rb failed ***
Could not create Makefile due to some reason, probably lack of
necessary libraries and/or headers. Check the mkmf.log file for more
details. You may need configuration options.
|
1.0
|
gem install pg on windows 7 fails - **[Original report](https://bitbucket.org/ged/ruby-pg/issue/64) by Anonymous.**
----------------------------------------
When I run "bundle install", I get this error:
checking for pg_config... yes
Using config values from c:\PostgreSQL\9.0\bin/pg_config.exe
checking for libpq-fe.h... yes
checking for libpq/libpq-fs.h... yes
checking for PQconnectdb() in -lpq... no
checking for PQconnectdb() in -llibpq... no
checking for PQconnectdb() in -lms/libpq... no
Can't find the PostgreSQL client library (libpq)
*** extconf.rb failed ***
Could not create Makefile due to some reason, probably lack of
necessary libraries and/or headers. Check the mkmf.log file for more
details. You may need configuration options.
|
non_process
|
gem install pg on windows fails by anonymous when i run bundle install i get this error checking for pg config yes using config values from c postgresql bin pg config exe checking for libpq fe h yes checking for libpq libpq fs h yes checking for pqconnectdb in lpq no checking for pqconnectdb in llibpq no checking for pqconnectdb in lms libpq no can t find the postgresql client library libpq extconf rb failed could not create makefile due to some reason probably lack of necessary libraries and or headers check the mkmf log file for more details you may need configuration options
| 0
|
184,588
| 14,289,501,639
|
IssuesEvent
|
2020-11-23 19:21:45
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
zhanshicai/zhanshicai: gossip/gossip/orgs_test.go; 14 LoC
|
fresh small test
|
Found a possible issue in [zhanshicai/zhanshicai](https://www.github.com/zhanshicai/zhanshicai) at [gossip/gossip/orgs_test.go](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ch := range channels {
if isOrgInChan(org, ch) {
for _, p := range peers {
p.JoinChan(joinChanMsgsByChan[ch], common.ChainID(ch))
p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch))
go func(p Gossip) {
for i := 0; i < 5; i++ {
time.Sleep(time.Second)
p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch))
}
}(p)
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096
|
1.0
|
zhanshicai/zhanshicai: gossip/gossip/orgs_test.go; 14 LoC -
Found a possible issue in [zhanshicai/zhanshicai](https://www.github.com/zhanshicai/zhanshicai) at [gossip/gossip/orgs_test.go](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/zhanshicai/zhanshicai/blob/fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096/gossip/gossip/orgs_test.go#L397-L410)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ch := range channels {
if isOrgInChan(org, ch) {
for _, p := range peers {
p.JoinChan(joinChanMsgsByChan[ch], common.ChainID(ch))
p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch))
go func(p Gossip) {
for i := 0; i < 5; i++ {
time.Sleep(time.Second)
p.UpdateChannelMetadata(createMetadata(1), common.ChainID(ch))
}
}(p)
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: fc4298bdbe0e2f79f8a3a80c31ba0ac46dc91096
|
non_process
|
zhanshicai zhanshicai gossip gossip orgs test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for ch range channels if isorginchan org ch for p range peers p joinchan joinchanmsgsbychan common chainid ch p updatechannelmetadata createmetadata common chainid ch go func p gossip for i i i time sleep time second p updatechannelmetadata createmetadata common chainid ch p leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
359,175
| 25,224,407,260
|
IssuesEvent
|
2022-11-14 15:01:57
|
stratosphereips/nist-cve-search-tool
|
https://api.github.com/repos/stratosphereips/nist-cve-search-tool
|
closed
|
Add documentation on how to use the Docker image
|
documentation
|
There's now a docker image, mostly for ensuring LTS. Add to the readme that the docker image exists, how to get it and a few examples of how to use it.
|
1.0
|
Add documentation on how to use the Docker image - There's now a docker image, mostly for ensuring LTS. Add to the readme that the docker image exists, how to get it and a few examples of how to use it.
|
non_process
|
add documentation on how to use the docker image there s now a docker image mostly for ensuring lts add to the readme that the docker image exists how to get it and a few examples of how to use it
| 0
|
452,730
| 32,066,489,936
|
IssuesEvent
|
2023-09-25 03:45:46
|
apache/incubator-opendal
|
https://api.github.com/repos/apache/incubator-opendal
|
closed
|
docs: Update the announcement email template to ensure that disclaimers have been added
|
documentation good first issue help wanted
|
As mentioned in the discussion at https://lists.apache.org/thread/0d8dxxbhb28m7w7kqxs1d2n0hsdybocg, Sebb reminded us to include disclaimers in our announcement.
We should add the following content:
```
---
Apache OpenDAL (incubating) is an effort undergoing incubation at the Apache
Software Foundation (ASF), sponsored by the Apache Incubator PMC.
Incubation is required of all newly accepted projects until a further review
indicates that the infrastructure, communications, and decision making process
have stabilized in a manner consistent with other successful ASF projects.
While incubation status is not necessarily a reflection of the completeness
or stability of the code, it does indicate that the project has yet to be
fully endorsed by the ASF.
```
|
1.0
|
docs: Update the announcement email template to ensure that disclaimers have been added - As mentioned in the discussion at https://lists.apache.org/thread/0d8dxxbhb28m7w7kqxs1d2n0hsdybocg, Sebb reminded us to include disclaimers in our announcement.
We should add the following content:
```
---
Apache OpenDAL (incubating) is an effort undergoing incubation at the Apache
Software Foundation (ASF), sponsored by the Apache Incubator PMC.
Incubation is required of all newly accepted projects until a further review
indicates that the infrastructure, communications, and decision making process
have stabilized in a manner consistent with other successful ASF projects.
While incubation status is not necessarily a reflection of the completeness
or stability of the code, it does indicate that the project has yet to be
fully endorsed by the ASF.
```
|
non_process
|
docs update the announcement email template to ensure that disclaimers have been added as mentioned in the discussion at sebb reminded us to include disclaimers in our announcement we should add the following content apache opendal incubating is an effort undergoing incubation at the apache software foundation asf sponsored by the apache incubator pmc incubation is required of all newly accepted projects until a further review indicates that the infrastructure communications and decision making process have stabilized in a manner consistent with other successful asf projects while incubation status is not necessarily a reflection of the completeness or stability of the code it does indicate that the project has yet to be fully endorsed by the asf
| 0
|
21,315
| 11,188,430,039
|
IssuesEvent
|
2020-01-02 04:57:52
|
0xProject/OpenZKP
|
https://api.github.com/repos/0xProject/OpenZKP
|
closed
|
Special case z == FieldElement::ONE?
|
performance tracker
|
*On 2019-04-23 @Recmo wrote in [`87f22ab`](https://github.com/0xProject/OpenZKP/commit/87f22ab866dbb5241a13c5916a726dd6047ed33d) “Implement edge cases in Jacobian”:*
Special case z == FieldElement::ONE?
See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl
```rust
self.x = x.clone();
self.y = y.clone();
self.z = FieldElement::ONE;
return;
}
// OPT: Special case z == FieldElement::ONE?
// See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl
let z1z1 = self.z.square();
let u2 = x * &z1z1;
let s2 = y * &self.z * &z1z1;
if self.x == u2 {
return if self.x == s2 {
```
*From [`algebra/elliptic-curve/src/jacobian.rs:199`](https://github.com/0xProject/OpenZKP/blob/3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2/algebra/elliptic-curve/src/jacobian.rs#L199)*
<!--{"commit-hash": "87f22ab866dbb5241a13c5916a726dd6047ed33d", "author": "Remco Bloemen", "author-mail": "<remco@0x.org>", "author-time": 1556062052, "author-tz": "-0700", "committer": "Remco Bloemen", "committer-mail": "<remco@0x.org>", "committer-time": 1556062052, "committer-tz": "-0700", "summary": "Implement edge cases in Jacobian", "previous": "77dc5773867b7ce66823b331137a1e64b353eb1c src/jacobian.rs", "filename": "algebra/elliptic-curve/src/jacobian.rs", "line": 198, "line_end": 200, "kind": "OPT", "issue": "Special case z == FieldElement::ONE?\nSee http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl", "head": "Special case z == FieldElement::ONE?", "context": " self.x = x.clone();\n self.y = y.clone();\n self.z = FieldElement::ONE;\n return;\n }\n // OPT: Special case z == FieldElement::ONE?\n // See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl\n let z1z1 = self.z.square();\n let u2 = x * &z1z1;\n let s2 = y * &self.z * &z1z1;\n if self.x == u2 {\n return if self.x == s2 {\n", "repo": "0xProject/OpenZKP", "branch-hash": "3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2"}-->
|
True
|
Special case z == FieldElement::ONE? - *On 2019-04-23 @Recmo wrote in [`87f22ab`](https://github.com/0xProject/OpenZKP/commit/87f22ab866dbb5241a13c5916a726dd6047ed33d) “Implement edge cases in Jacobian”:*
Special case z == FieldElement::ONE?
See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl
```rust
self.x = x.clone();
self.y = y.clone();
self.z = FieldElement::ONE;
return;
}
// OPT: Special case z == FieldElement::ONE?
// See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl
let z1z1 = self.z.square();
let u2 = x * &z1z1;
let s2 = y * &self.z * &z1z1;
if self.x == u2 {
return if self.x == s2 {
```
*From [`algebra/elliptic-curve/src/jacobian.rs:199`](https://github.com/0xProject/OpenZKP/blob/3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2/algebra/elliptic-curve/src/jacobian.rs#L199)*
<!--{"commit-hash": "87f22ab866dbb5241a13c5916a726dd6047ed33d", "author": "Remco Bloemen", "author-mail": "<remco@0x.org>", "author-time": 1556062052, "author-tz": "-0700", "committer": "Remco Bloemen", "committer-mail": "<remco@0x.org>", "committer-time": 1556062052, "committer-tz": "-0700", "summary": "Implement edge cases in Jacobian", "previous": "77dc5773867b7ce66823b331137a1e64b353eb1c src/jacobian.rs", "filename": "algebra/elliptic-curve/src/jacobian.rs", "line": 198, "line_end": 200, "kind": "OPT", "issue": "Special case z == FieldElement::ONE?\nSee http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl", "head": "Special case z == FieldElement::ONE?", "context": " self.x = x.clone();\n self.y = y.clone();\n self.z = FieldElement::ONE;\n return;\n }\n // OPT: Special case z == FieldElement::ONE?\n // See http://www.hyperelliptic.org/EFD/g1p/auto-shortw-jacobian.html#addition-madd-2007-bl\n let z1z1 = self.z.square();\n let u2 = x * &z1z1;\n let s2 = y * &self.z * &z1z1;\n if self.x == u2 {\n return if self.x == s2 {\n", "repo": "0xProject/OpenZKP", "branch-hash": "3fad461cf5bfd4cc08a4cc269a63ea2ac82861c2"}-->
|
non_process
|
special case z fieldelement one on recmo wrote in “implement edge cases in jacobian” special case z fieldelement one see rust self x x clone self y y clone self z fieldelement one return opt special case z fieldelement one see let self z square let x let y self z if self x return if self x from author time author tz committer remco bloemen committer mail committer time committer tz summary implement edge cases in jacobian previous src jacobian rs filename algebra elliptic curve src jacobian rs line line end kind opt issue special case z fieldelement one nsee head special case z fieldelement one context self x x clone n self y y clone n self z fieldelement one n return n n opt special case z fieldelement one n see let self z square n let x n let y self z n if self x n return if self x n repo openzkp branch hash
| 0
|
1,880
| 4,019,362,958
|
IssuesEvent
|
2016-05-16 14:41:41
|
psu-libraries/library_services_model
|
https://api.github.com/repos/psu-libraries/library_services_model
|
opened
|
Join everyone to the repo
|
services development
|
### Requirements
- [ ] Rama invited
- [ ] Stephen invited
- [ ] Patricia invited
- [ ] Nathan invited
- [ ] Robyn invited
|
1.0
|
Join everyone to the repo - ### Requirements
- [ ] Rama invited
- [ ] Stephen invited
- [ ] Patricia invited
- [ ] Nathan invited
- [ ] Robyn invited
|
non_process
|
join everyone to the repo requirements rama invited stephen invited patricia invited nathan invited robyn invited
| 0
|
12,357
| 14,887,185,073
|
IssuesEvent
|
2021-01-20 17:58:12
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Open study > Enrollment target > Entered data is getting removed
|
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
AR : Open study > Enrollment target > Entered data is getting removed when user enters the 6th digit
ER : It should not allow the user to enter 6th digit and entered data should not be removed
[ Note : It should also be fixed when user enters other characters after numbers ]
https://user-images.githubusercontent.com/60386291/103529284-65400280-4eab-11eb-82e6-015be0401bd1.mp4
|
3.0
|
[PM] Open study > Enrollment target > Entered data is getting removed - AR : Open study > Enrollment target > Entered data is getting removed when user enters the 6th digit
ER : It should not allow the user to enter 6th digit and entered data should not be removed
[ Note : It should also be fixed when user enters other characters after numbers ]
https://user-images.githubusercontent.com/60386291/103529284-65400280-4eab-11eb-82e6-015be0401bd1.mp4
|
process
|
open study enrollment target entered data is getting removed ar open study enrollment target entered data is getting removed when user enters the digit er it should not allow the user to enter digit and entered data should not be removed
| 1
|
271,851
| 20,719,592,014
|
IssuesEvent
|
2022-03-13 06:52:02
|
christian-cahig/Masterarbeit-APF
|
https://api.github.com/repos/christian-cahig/Masterarbeit-APF
|
opened
|
Better naming and notation for "generalized branch connection matrix" and "intermediate state vector"
|
documentation enhancement PyAPF APF.m
|
To be consistent with the "full" and "reduced" versions of some vectors and matrices, there should be a _full_ and a _reduced intermediate state vectors_ denoted by `$ \boldsymbol{x} $` (`x`) and `$ \boldsymbol{w} $` (`w`), respectively.
Moreover, it is a bit misleading to use the term "generalized branch connection matrix" when all that has been done is augment the directed and undirected branch-to-bus incidence matrices. It seems more accurate to therefore use _augmented branch-to-bus incidence matrix_. This will come in two variants: _full_ (`$ \boldsymbol{C} $` and `C`) and _reduced_ (`$ \boldsymbol{U} $` and `U`). Hence, the factorized formulation of the power flow equations will still consist of `$ \boldsymbol{E} \boldsymbol{y} = \boldsymbol{e} $` and `$ \boldsymbol{C} \boldsymbol{x} = \boldsymbol{u} $`, but the ones used in solving for the anticipated voltages are `$ \boldsymbol{Z} \boldsymbol{y} = \boldsymbol{z} $` and `$ \boldsymbol{U} \boldsymbol{w} = \boldsymbol{u} $`.
Consequently, the matrix currently denoted by `$ \boldsymbol{U} $` (`U`) needs to be assigned another symbol; `$ \boldsymbol{F} $` (`F`) seems fine.
|
1.0
|
Better naming and notation for "generalized branch connection matrix" and "intermediate state vector" - To be consistent with the "full" and "reduced" versions of some vectors and matrices, there should be a _full_ and a _reduced intermediate state vectors_ denoted by `$ \boldsymbol{x} $` (`x`) and `$ \boldsymbol{w} $` (`w`), respectively.
Moreover, it is a bit misleading to use the term "generalized branch connection matrix" when all that has been done is augment the directed and undirected branch-to-bus incidence matrices. It seems more accurate to therefore use _augmented branch-to-bus incidence matrix_. This will come in two variants: _full_ (`$ \boldsymbol{C} $` and `C`) and _reduced_ (`$ \boldsymbol{U} $` and `U`). Hence, the factorized formulation of the power flow equations will still consist of `$ \boldsymbol{E} \boldsymbol{y} = \boldsymbol{e} $` and `$ \boldsymbol{C} \boldsymbol{x} = \boldsymbol{u} $`, but the ones used in solving for the anticipated voltages are `$ \boldsymbol{Z} \boldsymbol{y} = \boldsymbol{z} $` and `$ \boldsymbol{U} \boldsymbol{w} = \boldsymbol{u} $`.
Consequently, the matrix currently denoted by `$ \boldsymbol{U} $` (`U`) needs to be assigned another symbol; `$ \boldsymbol{F} $` (`F`) seems fine.
|
non_process
|
better naming and notation for generalized branch connection matrix and intermediate state vector to be consistent with the full and reduced versions of some vectors and matrices there should be a full and a reduced intermediate state vectors denoted by boldsymbol x x and boldsymbol w w respectively moreover it is a bit misleading to use the term generalized branch connection matrix when all that has been done is augment the directed and undirected branch to bus incidence matrices it seems more accurate to therefore use augmented branch to bus incidence matrix this will come in two variants full boldsymbol c and c and reduced boldsymbol u and u hence the factorized formulation of the power flow equations will still consist of boldsymbol e boldsymbol y boldsymbol e and boldsymbol c boldsymbol x boldsymbol u but the ones used in solving for the anticipated voltages are boldsymbol z boldsymbol y boldsymbol z and boldsymbol u boldsymbol w boldsymbol u consequently the matrix currently denoted by boldsymbol u u needs to be assigned another symbol boldsymbol f f seems fine
| 0
|
22,146
| 7,124,795,382
|
IssuesEvent
|
2018-01-19 20:16:27
|
linode/manager
|
https://api.github.com/repos/linode/manager
|
closed
|
Upgrade to React 16
|
Backlog Build & Organization
|
This builds off of #2676
React 16 provides fibers, which should offer performance updates to our app, error boundaries, and a smaller footprint:
- https://edgecoders.com/react-16-features-and-fiber-explanation-e779544bb1b7
- https://reactjs.org/blog/2017/09/26/react-v16.0.html
- https://medium.com/netscape/whats-new-in-react-16-1608390ffe39
I took a stab at this but found that some of our external components have not been updated for React 16:
- https://github.com/akiran/react-highlight/pull/43
- https://github.com/HubSpot/react-select-plus/issues/94
- a few others
We should pursue those PRs (some are very stale) or replace those components with more active alternatives.
|
1.0
|
Upgrade to React 16 - This builds off of #2676
React 16 provides fibers, which should offer performance updates to our app, error boundaries, and a smaller footprint:
- https://edgecoders.com/react-16-features-and-fiber-explanation-e779544bb1b7
- https://reactjs.org/blog/2017/09/26/react-v16.0.html
- https://medium.com/netscape/whats-new-in-react-16-1608390ffe39
I took a stab at this but found that some of our external components have not been updated for React 16:
- https://github.com/akiran/react-highlight/pull/43
- https://github.com/HubSpot/react-select-plus/issues/94
- a few others
We should pursue those PRs (some are very stale) or replace those components with more active alternatives.
|
non_process
|
upgrade to react this builds off of react provides fibers which should offer performance updates to our app error boundaries and a smaller footprint i took a stab at this but found that some of our external components have not been updated for react a few others we should pursue those prs some are very stale or replace those components with more active alternatives
| 0
|
22,169
| 30,719,861,849
|
IssuesEvent
|
2023-07-27 15:13:22
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
@saltcorn/mobile-builder 0.8.7 has 3 guarddog issues
|
npm-install-script npm-silent-process-execution
|
```{"npm-install-script":[{"code":" \"postinstall\": \"node ./docker/post-installer.js\"","location":"package/package.json:10","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/dist/docker/post-installer.js:25","message":"This package is silently executing another executable"},{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/docker/post-installer.js:22","message":"This package is silently executing another executable"}]}```
|
1.0
|
@saltcorn/mobile-builder 0.8.7 has 3 guarddog issues - ```{"npm-install-script":[{"code":" \"postinstall\": \"node ./docker/post-installer.js\"","location":"package/package.json:10","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/dist/docker/post-installer.js:25","message":"This package is silently executing another executable"},{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/docker/post-installer.js:22","message":"This package is silently executing another executable"}]}```
|
process
|
saltcorn mobile builder has guarddog issues npm install script npm silent process execution
| 1
|
17,847
| 23,785,049,575
|
IssuesEvent
|
2022-09-02 09:18:09
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Improve logs in basic error reports
|
process/candidate kind/improvement tech/typescript topic: error reporting team/client
|
In many crash reports the logs look like this:
```
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
```
Example: https://github.com/prisma/prisma/issues/14248
We want to have more useful information here and probably don't need to log `libraryStarted` each time.
|
1.0
|
Improve logs in basic error reports - In many crash reports the logs look like this:
```
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
prisma:client:libraryEngine sending request, this.libraryStarted: true
```
Example: https://github.com/prisma/prisma/issues/14248
We want to have more useful information here and probably don't need to log `libraryStarted` each time.
|
process
|
improve logs in basic error reports in many crash reports the logs look like this prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true prisma client libraryengine sending request this librarystarted true example we want to have more useful information here and probably don t need to log librarystarted each time
| 1
|
5,371
| 8,202,228,676
|
IssuesEvent
|
2018-09-02 06:12:52
|
bio-miga/miga
|
https://api.github.com/repos/bio-miga/miga
|
closed
|
Re-registration of previous results not working
|
API Processing bug
|
Some steps involve (un)zipping files from previous steps, and registering the result should trigger a re-registration. However, the presence of the previous .json file stops this from happening.
Also, this re-registration should be moved from the `miga-base` Ruby code to the execution Bash scripts. It may need a new `--force` option for `miga add_result`.
|
1.0
|
Re-registration of previous results not working - Some steps involve (un)zipping files from previous steps, and registering the result should trigger a re-registration. However, the presence of the previous .json file stops this from happening.
Also, this re-registration should be moved from the `miga-base` Ruby code to the execution Bash scripts. It may need a new `--force` option for `miga add_result`.
|
process
|
re registration of previous results not working some steps involve un zipping files from previous steps and registering the result should trigger a re registration however the presence of the previous json file stops this from happening also this re registration should be moved from the miga base ruby code to the execution bash scripts it may need a new force option for miga add result
| 1
|
21,406
| 11,660,229,153
|
IssuesEvent
|
2020-03-03 02:37:44
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
opened
|
Feature: Unpaid Fee Flag on Permit Issuance
|
Product: AMANDA Project: ATD AMANDA Backlog Service: Apps Type: Enhancement Workgroup: ROW migrated
|
## Unpaid Fee Flag
_AKA # 46 Unpaid Fee Flag on Permit Issuance (Audit Request)_
**Stakeholders:** Paloma, Kim, Mooney
**User Story:** As a ROW permit analyst (PLA), I need to know if any of the folder people have unpaid ROW fees BEFORE the permit is issued, so I can place the permit on hold and attempt to collect the outstanding fee(s).
**Current State:**
1. The PLA, prior to permit approval, performs manual research to identify any unpaid ROW fees for ALL people listed on the people tab (Contractor, applicant, contact, whoever).
- They do not collect fees billed before 10/1/2013.
2. If no outstanding fees are identified, the permit is issued. End process.
3. If outstanding ATD fees are identified, the PLA provides the outstanding fee information to the person/entity responsible for payment.
- So if there are 3 unique people/entities on the permit, each entity will be provided with their unique fee information.
- The people/entities are provided with instructions for paying, all outstanding bill numbers, and the total for each bill number.
4. The PLA adds a note to the folder description: "Pending Outstanding Fee Payment **TL"
5. The customer will pay the fees, either online or in person.
6. The customer notifies PLA of payment
- When this step is missed, it leads to customers working in the ROW without an approved permit
7. PLA issues the permit. End Process.
**Future State:**
1. Folder issuance is attempted, either by a user from the folder tab, or by a system procedure like post payment.
- We need to be sure any new procedure doesn't cause payment update failures when the fees are paid online. (Put simply, lets learn from the issues caused by the Post Payment upgrade.)
- This feature may have to be split into 2 features, one for running this procedure when issuance is initiated by the user from the folder tab, and another for running this procedure when issuance is initiated by a post payment procedure. **Will need LIT feedback on this.**
2. The system performs the unpaid fee verification
- See # 1 in Current State; the system needs to do this instead of the PLA
- Query all unpaid fees for each people record on the people tab. Of those fees, exclude any not in folder types RW, EX, LC, ECV, GF, or DS. Of those fees, exclude any created prior to 10/1/2013. **Now you have a list of qualifying fees.**
3. If **qualifying fees** are found, an alert window opens informing the PLA that unpaid fees were found.
- The window will display this text: A people record related to this permit has unpaid ROW fees!
- **Stakeholder input needed.** Please provide feedback on the proposed content of the alert window:

- The window will facilitate the following actions:
- Allow the user to click a button to continue with permit issuance
- Allow the user to click a button to abort the permit issuance, and view information related to the qualifying fees
- The people invoice list view is an example of how we might provide the user with data about the qualifying fees:

4. PLA will provide the unpaid fee information (see list view above) to the customer, and note 'Pending Outstanding Fee Payment**' in the folder description.
- If it's possible to email the 'list view' of unpaid fees to the customer from the window in AMANDA, that would be ideal **Will need LIT feedback on this.**
5. Customer pays
- Jump to # 8 if customer does not pay
6. AMANDA notifies the PLA (or perhaps a resource task list?) that the permit needs to be issued **Stakeholder input needed. How would they most like to be notified?**
7. PLA issues the permit. End Process.
8. **Stakeholder input needed.** Will ROW need/want to know that these fees are still unpaid after a certain amount of time? For example, a monthly report emailed by the system with info about all 'fee flag alerts' that occurred in the previous month; only include info for fees that are still unpaid.
- This would be another story/feature however LIT needs to be aware of the reporting requirement when developing the fee flag feature.
------
Old Description: Requesting an alert to notify analysts when there are unpaid fees affiliated with any people records tied to a folder. Something similar to the current alert received when an analyst attempts to issue a permit that has violations affiliated with an address records tied to a folder.
Request Date: 2018-11-29 17:12:00
Request ID: DTS18-103080
Status: Backlog
Customer Priority: 0
Level of Effort: Small Project
DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c00727bd1bf2c4a9f4d268c
*Migrated from [atd-amanda #46](https://github.com/cityofaustin/atd-amanda/issues/46)*
|
1.0
|
Feature: Unpaid Fee Flag on Permit Issuance - ## Unpaid Fee Flag
_AKA # 46 Unpaid Fee Flag on Permit Issuance (Audit Request)_
**Stakeholders:** Paloma, Kim, Mooney
**User Story:** As a ROW permit analyst (PLA), I need to know if any of the folder people have unpaid ROW fees BEFORE the permit is issued, so I can place the permit on hold and attempt to collect the outstanding fee(s).
**Current State:**
1. The PLA, prior to permit approval, performs manual research to identify any unpaid ROW fees for ALL people listed on the people tab (Contractor, applicant, contact, whoever).
- They do not collect fees billed before 10/1/2013.
2. If no outstanding fees are identified, the permit is issued. End process.
3. If outstanding ATD fees are identified, the PLA provides the outstanding fee information to the person/entity responsible for payment.
- So if there are 3 unique people/entities on the permit, each entity will be provided with their unique fee information.
- The people/entities are provided with instructions for paying, all outstanding bill numbers, and the total for each bill number.
4. The PLA adds a note to the folder description: "Pending Outstanding Fee Payment **TL"
5. The customer will pay the fees, either online or in person.
6. The customer notifies PLA of payment
- When this step is missed, it leads to customers working in the ROW without an approved permit
7. PLA issues the permit. End Process.
**Future State:**
1. Folder issuance is attempted, either by a user from the folder tab, or by a system procedure like post payment.
- We need to be sure any new procedure doesn't cause payment update failures when the fees are paid online. (Put simply, lets learn from the issues caused by the Post Payment upgrade.)
- This feature may have to be split into 2 features, one for running this procedure when issuance is initiated by the user from the folder tab, and another for running this procedure when issuance is initiated by a post payment procedure. **Will need LIT feedback on this.**
2. The system performs the unpaid fee verification
- See # 1 in Current State; the system needs to do this instead of the PLA
- Query all unpaid fees for each people record on the people tab. Of those fees, exclude any not in folder types RW, EX, LC, ECV, GF, or DS. Of those fees, exclude any created prior to 10/1/2013. **Now you have a list of qualifying fees.**
3. If **qualifying fees** are found, an alert window opens informing the PLA that unpaid fees were found.
- The window will display this text: A people record related to this permit has unpaid ROW fees!
- **Stakeholder input needed.** Please provide feedback on the proposed content of the alert window:

- The window will facilitate the following actions:
- Allow the user to click a button to continue with permit issuance
- Allow the user to click a button to abort the permit issuance, and view information related to the qualifying fees
- The people invoice list view is an example of how we might provide the user with data about the qualifying fees:

4. PLA will provide the unpaid fee information (see list view above) to the customer, and note 'Pending Outstanding Fee Payment**' in the folder description.
- If it's possible to email the 'list view' of unpaid fees to the customer from the window in AMANDA, that would be ideal **Will need LIT feedback on this.**
5. Customer pays
- Jump to # 8 if customer does not pay
6. AMANDA notifies the PLA (or perhaps a resource task list?) that the permit needs to be issued **Stakeholder input needed. How would they most like to be notified?**
7. PLA issues the permit. End Process.
8. **Stakeholder input needed.** Will ROW need/want to know that these fees are still unpaid after a certain amount of time? For example, a monthly report emailed by the system with info about all 'fee flag alerts' that occurred in the previous month; only include info for fees that are still unpaid.
- This would be another story/feature however LIT needs to be aware of the reporting requirement when developing the fee flag feature.
------
Old Description: Requesting an alert to notify analysts when there are unpaid fees affiliated with any people records tied to a folder. Something similar to the current alert received when an analyst attempts to issue a permit that has violations affiliated with an address records tied to a folder.
Request Date: 2018-11-29 17:12:00
Request ID: DTS18-103080
Status: Backlog
Customer Priority: 0
Level of Effort: Small Project
DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c00727bd1bf2c4a9f4d268c
*Migrated from [atd-amanda #46](https://github.com/cityofaustin/atd-amanda/issues/46)*
|
non_process
|
feature unpaid fee flag on permit issuance unpaid fee flag aka unpaid fee flag on permit issuance audit request stakeholders paloma kim mooney user story as a row permit analyst pla i need to know if any of the folder people have unpaid row fees before the permit is issued so i can place the permit on hold and attempt to collect the outstanding fee s current state the pla prior to permit approval performs manual research to identify any unpaid row fees for all people listed on the people tab contractor applicant contact whoever they do not collect fees billed before if no outstanding fees are identified the permit is issued end process if outstanding atd fees are identified the pla provides the outstanding fee information to the person entity responsible for payment so if there are unique people entities on the permit each entity will be provided with their unique fee information the people entities are provided with instructions for paying all outstanding bill numbers and the total for each bill number the pla adds a note to the folder description pending outstanding fee payment tl the customer will pay the fees either online or in person the customer notifies pla of payment when this step is missed it leads to customers working in the row without an approved permit pla issues the permit end process future state folder issuance is attempted either by a user from the folder tab or by a system procedure like post payment we need to be sure any new procedure doesn t cause payment update failures when the fees are paid online put simply lets learn from the issues caused by the post payment upgrade this feature may have to be split into features one for running this procedure when issuance is initiated by the user from the folder tab and another for running this procedure when issuance is initiated by a post payment procedure will need lit feedback on this the system performs the unpaid fee verification see in current state the system needs to do this instead of the pla query all unpaid fees for each people record on the people tab of those fees exclude any not in folder types rw ex lc ecv gf or ds of those fees exclude any created prior to now you have a list of qualifying fees if qualifying fees are found an alert window opens informing the pla that unpaid fees were found the window will display this text a people record related to this permit has unpaid row fees stakeholder input needed please provide feedback on the proposed content of the alert window the window will facilitate the following actions allow the user to click a button to continue with permit issuance allow the user to click a button to abort the permit issuance and view information related to the qualifying fees the people invoice list view is an example of how we might provide the user with data about the qualifying fees pla will provide the unpaid fee information see list view above to the customer and note pending outstanding fee payment in the folder description if it s possible to email the list view of unpaid fees to the customer from the window in amanda that would be ideal will need lit feedback on this customer pays jump to if customer does not pay amanda notifies the pla or perhaps a resource task list that the permit needs to be issued stakeholder input needed how would they most like to be notified pla issues the permit end process stakeholder input needed will row need want to know that these fees are still unpaid after a certain amount of time for example a monthly report emailed by the system with info about all fee flag alerts that occurred in the previous month only include info for fees that are still unpaid this would be another story feature however lit needs to be aware of the reporting requirement when developing the fee flag feature old description requesting an alert to notify analysts when there are unpaid fees affiliated with any people records tied to a folder something similar to the current alert received when an analyst attempts to issue a permit that has violations affiliated with an address records tied to a folder request date request id status backlog customer priority level of effort small project dts url migrated from
| 0
|
21,365
| 29,194,080,353
|
IssuesEvent
|
2023-05-20 00:31:48
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Hibrido / Caieiras, São Paulo, Brazil] C# Developer (Júnior) (Híbrido) na Coodesh
|
SALVADOR PJ BANCO DE DADOS FULL-STACK HTML JUNIOR SQL REST SOAP JSON ANGULAR REACT XML REQUISITOS PROCESSOS GITHUB UMA C APIs AUTOMAÇÃO DE PROCESSOS ALOCADO Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong><ins>TRIFORSYS</ins></strong> está em busca de <strong><ins>C# Developer (Júnior)</ins></strong> para integrar seu time!</p>
<p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p>
<p><strong>Sobre a empresa</strong></p>
<p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p>
<p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p>
<p></p>
<p>Venha com a gente fazer seu negócio!</p>
## Triforsys:
<p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p>
<p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a>
## Habilidades:
- Angular
- HTML
- CSS
- React.js
- Banco de dados relacionais (SQL)
## Local:
Caieiras, São Paulo, Brazil
## Requisitos:
- Residir em Caieiras ou proximidades (Formato híbrido);
- Conhecimentos em APIs, REST, SOAP;
- Conhecimentos em XML, JSON, HTML.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C# Developer (Júnior) (Híbrido) na Triforsys](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
PJ
#### Categoria
Full-Stack
|
2.0
|
[Hibrido / Caieiras, São Paulo, Brazil] C# Developer (Júnior) (Híbrido) na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong><ins>TRIFORSYS</ins></strong> está em busca de <strong><ins>C# Developer (Júnior)</ins></strong> para integrar seu time!</p>
<p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p>
<p><strong>Sobre a empresa</strong></p>
<p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p>
<p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p>
<p></p>
<p>Venha com a gente fazer seu negócio!</p>
## Triforsys:
<p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p>
<p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a>
## Habilidades:
- Angular
- HTML
- CSS
- React.js
- Banco de dados relacionais (SQL)
## Local:
Caieiras, São Paulo, Brazil
## Requisitos:
- Residir em Caieiras ou proximidades (Formato híbrido);
- Conhecimentos em APIs, REST, SOAP;
- Conhecimentos em XML, JSON, HTML.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C# Developer (Júnior) (Híbrido) na Triforsys](https://coodesh.com/vagas/analista-de-sistemas-junior-172722002?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
PJ
#### Categoria
Full-Stack
|
process
|
c developer júnior híbrido na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a triforsys está em busca de c developer júnior para integrar seu time fábrica softwares e soluções em ti de forma personalizada atendendo a necessidade do nosso cliente com foco na automação de processos sobre a empresa somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é venha com a gente fazer seu negócio triforsys somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é habilidades angular html css react js banco de dados relacionais sql local caieiras são paulo brazil requisitos residir em caieiras ou proximidades formato híbrido conhecimentos em apis rest soap conhecimentos em xml json html como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria full stack
| 1
|
20,426
| 3,812,781,005
|
IssuesEvent
|
2016-03-27 20:52:32
|
briansmith/ring
|
https://api.github.com/repos/briansmith/ring
|
closed
|
Remove all code that supports AES-192
|
enhancement good-first-bug performance static-analysis-and-type-safety test-coverage
|
See https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551
and
https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551
Also, I guess the assembly language code must have some bits that support 192-bit AES.
We don't need any of this since we don't (and don't plan to) expose AES-192.
|
1.0
|
Remove all code that supports AES-192 - See https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551
and
https://github.com/briansmith/ring/blob/c882c2c3ec837f14267991122dc20e6aba5700bc/crypto/aes/aes.c#L549-L551
Also, I guess the assembly language code must have some bits that support 192-bit AES.
We don't need any of this since we don't (and don't plan to) expose AES-192.
|
non_process
|
remove all code that supports aes see and also i guess the assembly language code must have some bits that support bit aes we don t need any of this since we don t and don t plan to expose aes
| 0
|
6,147
| 2,814,174,529
|
IssuesEvent
|
2015-05-18 18:32:45
|
joyent/node
|
https://api.github.com/repos/joyent/node
|
closed
|
test: gc/test-net-timeout fails consistently
|
test
|
```
Done: 500/500
Collected: 499/500
All should be collected now.
Collected: 499/500
timers.js:102
if (!process.listeners('uncaughtException').length) throw e;
^
AssertionError: false == true
at null._onTimeout (/home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js:64:7)
at Timer.list.ontimeout (timers.js:100:19)
at process._makeCallback (node.js:248:20)
Command: out/Release/node --expose-gc /home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js
```
/cc @isaacs?
|
1.0
|
test: gc/test-net-timeout fails consistently - ```
Done: 500/500
Collected: 499/500
All should be collected now.
Collected: 499/500
timers.js:102
if (!process.listeners('uncaughtException').length) throw e;
^
AssertionError: false == true
at null._onTimeout (/home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js:64:7)
at Timer.list.ontimeout (timers.js:100:19)
at process._makeCallback (node.js:248:20)
Command: out/Release/node --expose-gc /home/bnoordhuis/src/nodejs/master/test/gc/test-net-timeout.js
```
/cc @isaacs?
|
non_process
|
test gc test net timeout fails consistently done collected all should be collected now collected timers js if process listeners uncaughtexception length throw e assertionerror false true at null ontimeout home bnoordhuis src nodejs master test gc test net timeout js at timer list ontimeout timers js at process makecallback node js command out release node expose gc home bnoordhuis src nodejs master test gc test net timeout js cc isaacs
| 0
|
371,868
| 10,982,278,821
|
IssuesEvent
|
2019-12-01 06:03:33
|
Luna-Interactive/catastrophe
|
https://api.github.com/repos/Luna-Interactive/catastrophe
|
closed
|
Winning
|
Priority: Critical SFX
|
**Is your feature request related to a problem? Please describe.**
Missing sound for when a player wins the game.
|
1.0
|
Winning - **Is your feature request related to a problem? Please describe.**
Missing sound for when a player wins the game.
|
non_process
|
winning is your feature request related to a problem please describe missing sound for when a player wins the game
| 0
|
5,241
| 8,036,963,753
|
IssuesEvent
|
2018-07-30 10:55:24
|
Open-EO/openeo-api
|
https://api.github.com/repos/Open-EO/openeo-api
|
closed
|
Variables in process_graphs?
|
process graph management processes vote work in progress
|
In issue #52 about product name differences, I had the idea that stored process graphs, i.e. /process_graphs/:id, may need some variables in them to be really portable. Otherwise sharing as proposed in #85 might not be so useful. An example would be the product id in case we can't come up with something useful to solve the product name differences. Process graphs wouldn't be portable in case a back-end specific product name would be stored with the process_graph.
Changes to `POST /process_graphs` (and similarly `PATCH /process_graphs/:id`) could look like this:
```
{
"process_graph":{
"process_id":"min_time",
"description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.",
"imagery":{
"process_id":"NDVI",
"imagery":{
"process_id":"filter_daterange",
"imagery":{
"process_id":"get_data",
"data_id":"{$product}"
},
"from":"2017-01-01",
"to":"2017-01-31"
},
"red":"{$red_band}",
"nir":"{$nir_band}"
}
},
"variables":{
"product":{
"type":"string",
"default":"Sentinel-2A",
"description":"Product id of the dataset"
},
"red_band":{
"type":"string",
"default":"B4",
"description":"Red Band id of the dataset"
},
"nir_band":{
"type":"string",
"default":"B8",
"description":"NIR Band id of the dataset"
}
}
}
```
`"{$product}"`, `"{$red_band}"` and `"{$nir_band}"` are the variables to be replaced - including the quotes to keep it being valid, parsable JSON.
We could an additional (but optional) property "variables" that could hold the data type, default values and a description.
Types could be: everything that is a valid JSON data type, including objects and arrays.
Alternatively, we could make it look like JSON schema:
```
{
"process_graph":{
"process_id":"min_time",
"description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.",
"imagery":{
"process_id":"NDVI",
"imagery":{
"process_id":"filter_daterange",
"imagery":{
"process_id":"get_data",
"data_id":{
"$ref":"product",
"type":"string",
"default":"Sentinel-2A",
"description":"Product id of the dataset"
}
},
"from":"2017-01-01",
"to":"2017-01-31"
},
"red":{
"$ref":"red_band",
"type":"string",
"default":"B4",
"description":"Red Band id of the dataset"
},
"nir":{
"$ref":"nir_band",
"type":"string",
"default":"B8",
"description":"NIR Band id of the dataset"
}
}
}
}
```
That approach allows for better encapsulation. The "$ref" key would be the "identifier" for the variables JSON object. "$ref" could also be "variable_id" to be more in-line with "process_id".
`GET /process_graphs/:id` could work in two modes:
1. Variables could be transferred **serialized** as JSON via query parameters and the back-end returns the resulting process graph
2. The requesting entities processes/replaces the variables.
Example for the first approach:
Example request: `GET https://www.openeo.org/api/v0/process_graphs/1234?product=%22Sentinel-2A-L1C%22`
The product would be changed and the bands would be taken from the specified defaults.
Resulting process_graph:
```
{
"process_id":"min_time",
"description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.",
"imagery":{
"process_id":"NDVI",
"imagery":{
"process_id":"filter_daterange",
"imagery":{
"process_id":"get_data",
"data_id":"Sentinal-2A-L1C"
},
"from":"2017-01-01",
"to":"2017-01-31"
},
"red":"B4",
"nir":"B8"
}
}
```
Both modes could be used in a process like this:
```
{
"process_id":"process_graph",
"url":"https://www.openeo.org/api/v0/process_graphs/1234",
"product":"Sentinal-2A-L1C"
}
```
|
2.0
|
Variables in process_graphs? - In issue #52 about product name differences, I had the idea that stored process graphs, i.e. /process_graphs/:id, may need some variables in them to be really portable. Otherwise sharing as proposed in #85 might not be so useful. An example would be the product id in case we can't come up with something useful to solve the product name differences. Process graphs wouldn't be portable in case a back-end specific product name would be stored with the process_graph.
Changes to `POST /process_graphs` (and similarly `PATCH /process_graphs/:id`) could look like this:
```
{
"process_graph":{
"process_id":"min_time",
"description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.",
"imagery":{
"process_id":"NDVI",
"imagery":{
"process_id":"filter_daterange",
"imagery":{
"process_id":"get_data",
"data_id":"{$product}"
},
"from":"2017-01-01",
"to":"2017-01-31"
},
"red":"{$red_band}",
"nir":"{$nir_band}"
}
},
"variables":{
"product":{
"type":"string",
"default":"Sentinel-2A",
"description":"Product id of the dataset"
},
"red_band":{
"type":"string",
"default":"B4",
"description":"Red Band id of the dataset"
},
"nir_band":{
"type":"string",
"default":"B8",
"description":"NIR Band id of the dataset"
}
}
}
```
`"{$product}"`, `"{$red_band}"` and `"{$nir_band}"` are the variables to be replaced - including the quotes to keep it being valid, parsable JSON.
We could an additional (but optional) property "variables" that could hold the data type, default values and a description.
Types could be: everything that is a valid JSON data type, including objects and arrays.
Alternatively, we could make it look like JSON schema:
```
{
"process_graph":{
"process_id":"min_time",
"description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.",
"imagery":{
"process_id":"NDVI",
"imagery":{
"process_id":"filter_daterange",
"imagery":{
"process_id":"get_data",
"data_id":{
"$ref":"product",
"type":"string",
"default":"Sentinel-2A",
"description":"Product id of the dataset"
}
},
"from":"2017-01-01",
"to":"2017-01-31"
},
"red":{
"$ref":"red_band",
"type":"string",
"default":"B4",
"description":"Red Band id of the dataset"
},
"nir":{
"$ref":"nir_band",
"type":"string",
"default":"B8",
"description":"NIR Band id of the dataset"
}
}
}
}
```
That approach allows for better encapsulation. The "$ref" key would be the "identifier" for the variables JSON object. "$ref" could also be "variable_id" to be more in-line with "process_id".
`GET /process_graphs/:id` could work in two modes:
1. Variables could be transferred **serialized** as JSON via query parameters and the back-end returns the resulting process graph
2. The requesting entities processes/replaces the variables.
Example for the first approach:
Example request: `GET https://www.openeo.org/api/v0/process_graphs/1234?product=%22Sentinel-2A-L1C%22`
The product would be changed and the bands would be taken from the specified defaults.
Resulting process_graph:
```
{
"process_id":"min_time",
"description":"Deriving minimum NDVI measurements over pixel time series of Sentinel 2 imagery.",
"imagery":{
"process_id":"NDVI",
"imagery":{
"process_id":"filter_daterange",
"imagery":{
"process_id":"get_data",
"data_id":"Sentinal-2A-L1C"
},
"from":"2017-01-01",
"to":"2017-01-31"
},
"red":"B4",
"nir":"B8"
}
}
```
Both modes could be used in a process like this:
```
{
"process_id":"process_graph",
"url":"https://www.openeo.org/api/v0/process_graphs/1234",
"product":"Sentinal-2A-L1C"
}
```
|
process
|
variables in process graphs in issue about product name differences i had the idea that stored process graphs i e process graphs id may need some variables in them to be really portable otherwise sharing as proposed in might not be so useful an example would be the product id in case we can t come up with something useful to solve the product name differences process graphs wouldn t be portable in case a back end specific product name would be stored with the process graph changes to post process graphs and similarly patch process graphs id could look like this process graph process id min time description deriving minimum ndvi measurements over pixel time series of sentinel imagery imagery process id ndvi imagery process id filter daterange imagery process id get data data id product from to red red band nir nir band variables product type string default sentinel description product id of the dataset red band type string default description red band id of the dataset nir band type string default description nir band id of the dataset product red band and nir band are the variables to be replaced including the quotes to keep it being valid parsable json we could an additional but optional property variables that could hold the data type default values and a description types could be everything that is a valid json data type including objects and arrays alternatively we could make it look like json schema process graph process id min time description deriving minimum ndvi measurements over pixel time series of sentinel imagery imagery process id ndvi imagery process id filter daterange imagery process id get data data id ref product type string default sentinel description product id of the dataset from to red ref red band type string default description red band id of the dataset nir ref nir band type string default description nir band id of the dataset that approach allows for better encapsulation the ref key would be the identifier for the variables json object ref could also be variable id to be more in line with process id get process graphs id could work in two modes variables could be transferred serialized as json via query parameters and the back end returns the resulting process graph the requesting entities processes replaces the variables example for the first approach example request get the product would be changed and the bands would be taken from the specified defaults resulting process graph process id min time description deriving minimum ndvi measurements over pixel time series of sentinel imagery imagery process id ndvi imagery process id filter daterange imagery process id get data data id sentinal from to red nir both modes could be used in a process like this process id process graph url product sentinal
| 1
|
10,118
| 13,044,162,241
|
IssuesEvent
|
2020-07-29 03:47:31
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `GetFormat` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `GetFormat` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `GetFormat` from TiDB -
## Description
Port the scalar function `GetFormat` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function getformat from tidb description port the scalar function getformat from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
13,102
| 15,496,467,659
|
IssuesEvent
|
2021-03-11 02:44:05
|
dluiscosta/weather_api
|
https://api.github.com/repos/dluiscosta/weather_api
|
opened
|
Establish test dependency hierarchy
|
development process enhancement
|
Establish test dependency hierarchy where some test ```A``` might assert the proper operation of a feature, which in turn requires the proper operation of a second feature, asserted independently by some test ```B```, thus making ```A``` dependent on ```B```.
That way, when ```B``` fails, ```A``` can be skipped (as it's bound to fail as well), thus saving time.
|
1.0
|
Establish test dependency hierarchy - Establish test dependency hierarchy where some test ```A``` might assert the proper operation of a feature, which in turn requires the proper operation of a second feature, asserted independently by some test ```B```, thus making ```A``` dependent on ```B```.
That way, when ```B``` fails, ```A``` can be skipped (as it's bound to fail as well), thus saving time.
|
process
|
establish test dependency hierarchy establish test dependency hierarchy where some test a might assert the proper operation of a feature which in turn requires the proper operation of a second feature asserted independently by some test b thus making a dependent on b that way when b fails a can be skipped as it s bound to fail as well thus saving time
| 1
|
18,067
| 24,080,244,816
|
IssuesEvent
|
2022-09-19 05:32:02
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Processing algorithm "Lines to polygons" produces incorrect polygons
|
Processing Bug
|
### What is the bug or the crash?
When using the processing algorithm "Lines to polygons" with certain data, it produces incorrect polygons.
Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip)
### Steps to reproduce the issue
1. Add lines_simplified.json to the project

2. Run "Lines to polygons" processing algorithm with the layer

### Versions
QGISin versio
3.22.9-Białowieża
QGIS-koodin versio
a8e9e6fa
Qt:n versio
5.15.3
Python-versio
3.9.5
GDAL/OGR:n versio
3.5.1
PROJ-versio
9.0.1
EPSG-rekisteritietokannan versio
v10.064 (2022-05-19)
GEOS-versio
3.10.3-CAPI-1.16.1
SQLite-versio
3.38.1
PDAL-versio
2.3.0
PostgreSQL:n asiakkaan versio
14.3
SpatiaLite-versio
5.0.1
QWT-versio
6.1.6
QScintilla2-versio
2.13.1
Käyttöjärjestelmän versio
Windows 10 Version 2009
Aktiiviset python-lisäosat
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
plugin_reloader
0.9.1
Ilre
0.1.0
Also tested with QGIS 3.26.2 on Arch Linux
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The problem might lie with the dataset used. The same dataset causes the problems in #50227
|
1.0
|
Processing algorithm "Lines to polygons" produces incorrect polygons - ### What is the bug or the crash?
When using the processing algorithm "Lines to polygons" with certain data, it produces incorrect polygons.
Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip)
### Steps to reproduce the issue
1. Add lines_simplified.json to the project

2. Run "Lines to polygons" processing algorithm with the layer

### Versions
QGISin versio
3.22.9-Białowieża
QGIS-koodin versio
a8e9e6fa
Qt:n versio
5.15.3
Python-versio
3.9.5
GDAL/OGR:n versio
3.5.1
PROJ-versio
9.0.1
EPSG-rekisteritietokannan versio
v10.064 (2022-05-19)
GEOS-versio
3.10.3-CAPI-1.16.1
SQLite-versio
3.38.1
PDAL-versio
2.3.0
PostgreSQL:n asiakkaan versio
14.3
SpatiaLite-versio
5.0.1
QWT-versio
6.1.6
QScintilla2-versio
2.13.1
Käyttöjärjestelmän versio
Windows 10 Version 2009
Aktiiviset python-lisäosat
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
plugin_reloader
0.9.1
Ilre
0.1.0
Also tested with QGIS 3.26.2 on Arch Linux
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The problem might lie with the dataset used. The same dataset causes the problems in #50227
|
process
|
processing algorithm lines to polygons produces incorrect polygons what is the bug or the crash when using the processing algorithm lines to polygons with certain data it produces incorrect polygons here is a simple topologically correct line layer to test the bug with steps to reproduce the issue add lines simplified json to the project run lines to polygons processing algorithm with the layer versions qgisin versio białowieża qgis koodin versio qt n versio python versio gdal ogr n versio proj versio epsg rekisteritietokannan versio geos versio capi sqlite versio pdal versio postgresql n asiakkaan versio spatialite versio qwt versio versio käyttöjärjestelmän versio windows version aktiiviset python lisäosat db manager grassprovider metasearch processing sagaprovider plugin reloader ilre also tested with qgis on arch linux supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context the problem might lie with the dataset used the same dataset causes the problems in
| 1
|
408,546
| 27,695,438,910
|
IssuesEvent
|
2023-03-14 01:33:12
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
[Gitbook Documentation] Complete 'Extracting Location Coordinates to add a Map Marker Map' article
|
Workgroup: DTS Service: Apps Type: Documentation Product: TDS Portal
|
As exemplified in TDS or S&M
|
1.0
|
[Gitbook Documentation] Complete 'Extracting Location Coordinates to add a Map Marker Map' article - As exemplified in TDS or S&M
|
non_process
|
complete extracting location coordinates to add a map marker map article as exemplified in tds or s m
| 0
|
19,615
| 25,970,594,587
|
IssuesEvent
|
2022-12-19 10:54:07
|
toggl/track-windows-feedback
|
https://api.github.com/repos/toggl/track-windows-feedback
|
closed
|
Is there a Beta "Release channel"?
|
solved processed
|
I see the option for changing to the Beta "Release channel" but when I select it and then close the window and re-open it switches back to the Stable channel. Is there a Beta channel anymore? I know in the past (old app) there was and I was on the Beta channel but not sure if now with new native app if maybe there is only a Stable channel and this dropdown should be ignored?
Thanks!

|
1.0
|
Is there a Beta "Release channel"? - I see the option for changing to the Beta "Release channel" but when I select it and then close the window and re-open it switches back to the Stable channel. Is there a Beta channel anymore? I know in the past (old app) there was and I was on the Beta channel but not sure if now with new native app if maybe there is only a Stable channel and this dropdown should be ignored?
Thanks!

|
process
|
is there a beta release channel i see the option for changing to the beta release channel but when i select it and then close the window and re open it switches back to the stable channel is there a beta channel anymore i know in the past old app there was and i was on the beta channel but not sure if now with new native app if maybe there is only a stable channel and this dropdown should be ignored thanks
| 1
|
37,342
| 15,262,770,210
|
IssuesEvent
|
2021-02-22 00:38:34
|
Azure/azure-sdk-for-net
|
https://api.github.com/repos/Azure/azure-sdk-for-net
|
closed
|
Error while using Azure.Messaging.ServiceBus library for Service bus trigger function
|
Client Functions Service Bus customer-reported needs-author-feedback question
|
Azure.Messaging.ServiceBus 7.1.0
Microsoft.Azure.WebJobs.Extensions.ServiceBus 4.2.1
Microsoft.NET.Sdk.Functions 3.0.11
Azure Cloud.
Windows/Linux
VS 2019
Hello
I have an Azure function, with a Service Bus trigger,
using Azure.Messaging.ServiceBus;
public static void Run(ServiceBusReceivedMessage myQueueItem, ILogger log, Microsoft.Azure.WebJobs.ExecutionContext context) {..}
Is that supported? The [documentation](https://docs.microsoft.com/en-us/azure/azure-functions/functions-bindings-service-bus-trigger?tabs=csharp#usage) says that it should be used Microsoft.Azure.ServiceBus libray, but I need to used the newest libray.
The code throws an error so it doesn't seem to be supported..
Binding parameters to complex objects (such as 'ServiceBusReceivedMessage') uses Json.NET serialization or XML object serialization. 1. If ContentType is 'application/json' deserialize as JSON 2. If ContentType is not 'application/json' attempt to deserialize using Message.GetBody, which will handle cases like XML object serialization 3. If this deserialization fails, do a final attempt at JSON deserialization to catch cases where the content type might be incorrect The JSON parser failed: Unexpected character encountered while parsing value: T. Path '', line 0, position 0.
Thanks.
|
1.0
|
Error while using Azure.Messaging.ServiceBus library for Service bus trigger function - Azure.Messaging.ServiceBus 7.1.0
Microsoft.Azure.WebJobs.Extensions.ServiceBus 4.2.1
Microsoft.NET.Sdk.Functions 3.0.11
Azure Cloud.
Windows/Linux
VS 2019
Hello
I have an Azure function, with a Service Bus trigger,
using Azure.Messaging.ServiceBus;
public static void Run(ServiceBusReceivedMessage myQueueItem, ILogger log, Microsoft.Azure.WebJobs.ExecutionContext context) {..}
Is that supported? The [documentation](https://docs.microsoft.com/en-us/azure/azure-functions/functions-bindings-service-bus-trigger?tabs=csharp#usage) says that it should be used Microsoft.Azure.ServiceBus libray, but I need to used the newest libray.
The code throws an error so it doesn't seem to be supported..
Binding parameters to complex objects (such as 'ServiceBusReceivedMessage') uses Json.NET serialization or XML object serialization. 1. If ContentType is 'application/json' deserialize as JSON 2. If ContentType is not 'application/json' attempt to deserialize using Message.GetBody, which will handle cases like XML object serialization 3. If this deserialization fails, do a final attempt at JSON deserialization to catch cases where the content type might be incorrect The JSON parser failed: Unexpected character encountered while parsing value: T. Path '', line 0, position 0.
Thanks.
|
non_process
|
error while using azure messaging servicebus library for service bus trigger function azure messaging servicebus microsoft azure webjobs extensions servicebus microsoft net sdk functions azure cloud windows linux vs hello i have an azure function with a service bus trigger using azure messaging servicebus public static void run servicebusreceivedmessage myqueueitem ilogger log microsoft azure webjobs executioncontext context is that supported the says that it should be used microsoft azure servicebus libray but i need to used the newest libray the code throws an error so it doesn t seem to be supported binding parameters to complex objects such as servicebusreceivedmessage uses json net serialization or xml object serialization if contenttype is application json deserialize as json if contenttype is not application json attempt to deserialize using message getbody which will handle cases like xml object serialization if this deserialization fails do a final attempt at json deserialization to catch cases where the content type might be incorrect the json parser failed unexpected character encountered while parsing value t path line position thanks
| 0
|
217,525
| 16,855,802,384
|
IssuesEvent
|
2021-06-21 06:27:36
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
raftstore::test_merge::test_node_merge_prerequisites_check failed
|
component/test-bench
|
raftstore::test_merge::test_node_merge_prerequisites_check
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/18842/display/redirect
|
1.0
|
raftstore::test_merge::test_node_merge_prerequisites_check failed - raftstore::test_merge::test_node_merge_prerequisites_check
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/18842/display/redirect
|
non_process
|
raftstore test merge test node merge prerequisites check failed raftstore test merge test node merge prerequisites check latest failed builds
| 0
|
719,588
| 24,764,745,723
|
IssuesEvent
|
2022-10-22 11:17:00
|
DISSINET/InkVisitor
|
https://api.github.com/repos/DISSINET/InkVisitor
|
closed
|
New reference; changing the parsing notation for references
|
data parsing priority
|
1. Please include a new col. in Actions and Concepts into the parsing: wordnet_sense_key.
2. I had an inconsistency in the notation of wordnet cols between Actions and Concepts ("special" vs. "reference"). Now unified to "reference".
3. In cols. to be parsed as "reference", I now added the Resource ID (R entity) to the row no. 3, and created those entities in the normal [list of Rs](https://docs.google.com/spreadsheets/d/13eVorFf7J9R8YzO7TmJRVLzIIwRJS737r7eFbH1boyE/edit#gid=583750775).
|
1.0
|
New reference; changing the parsing notation for references - 1. Please include a new col. in Actions and Concepts into the parsing: wordnet_sense_key.
2. I had an inconsistency in the notation of wordnet cols between Actions and Concepts ("special" vs. "reference"). Now unified to "reference".
3. In cols. to be parsed as "reference", I now added the Resource ID (R entity) to the row no. 3, and created those entities in the normal [list of Rs](https://docs.google.com/spreadsheets/d/13eVorFf7J9R8YzO7TmJRVLzIIwRJS737r7eFbH1boyE/edit#gid=583750775).
|
non_process
|
new reference changing the parsing notation for references please include a new col in actions and concepts into the parsing wordnet sense key i had an inconsistency in the notation of wordnet cols between actions and concepts special vs reference now unified to reference in cols to be parsed as reference i now added the resource id r entity to the row no and created those entities in the normal
| 0
|
113,790
| 4,569,032,192
|
IssuesEvent
|
2016-09-15 16:02:01
|
isobar-techchallenge/isobar-a9f8d858-fb06-4865-833f-f03fc2d03cd1
|
https://api.github.com/repos/isobar-techchallenge/isobar-a9f8d858-fb06-4865-833f-f03fc2d03cd1
|
opened
|
Target Button in Email to Small
|
priority:low
|
Within email to user the CTA button target space is to small. Need to make the target click area the full width of the button.
|
1.0
|
Target Button in Email to Small - Within email to user the CTA button target space is to small. Need to make the target click area the full width of the button.
|
non_process
|
target button in email to small within email to user the cta button target space is to small need to make the target click area the full width of the button
| 0
|
18,148
| 24,187,126,890
|
IssuesEvent
|
2022-09-23 14:12:03
|
apache/arrow-rs
|
https://api.github.com/repos/apache/arrow-rs
|
closed
|
Golang Integration Tests Failing
|
bug development-process
|
**Describe the bug**
<!--
A clear and concise description of what the bug is.
-->
The arrow integration tests are failing on master, https://github.com/apache/arrow-rs/runs/8249872730?check_suite_focus=true, this appears to be caused by https://github.com/apache/arrow/pull/14067 which bumps the minimum required Golang version to 1.17.
Unfortunately the version in `apache/arrow-dev:amd64-conda-integration` is 1.16
**To Reproduce**
<!--
Steps to reproduce the behavior:
-->
**Expected behavior**
<!--
A clear and concise description of what you expected to happen.
-->
**Additional context**
<!--
Add any other context about the problem here.
-->
|
1.0
|
Golang Integration Tests Failing - **Describe the bug**
<!--
A clear and concise description of what the bug is.
-->
The arrow integration tests are failing on master, https://github.com/apache/arrow-rs/runs/8249872730?check_suite_focus=true, this appears to be caused by https://github.com/apache/arrow/pull/14067 which bumps the minimum required Golang version to 1.17.
Unfortunately the version in `apache/arrow-dev:amd64-conda-integration` is 1.16
**To Reproduce**
<!--
Steps to reproduce the behavior:
-->
**Expected behavior**
<!--
A clear and concise description of what you expected to happen.
-->
**Additional context**
<!--
Add any other context about the problem here.
-->
|
process
|
golang integration tests failing describe the bug a clear and concise description of what the bug is the arrow integration tests are failing on master this appears to be caused by which bumps the minimum required golang version to unfortunately the version in apache arrow dev conda integration is to reproduce steps to reproduce the behavior expected behavior a clear and concise description of what you expected to happen additional context add any other context about the problem here
| 1
|
445,518
| 12,832,125,680
|
IssuesEvent
|
2020-07-07 07:04:31
|
Automattic/abacus
|
https://api.github.com/repos/Automattic/abacus
|
opened
|
Apply consistent object normalisation across the codebase
|
[!priority] medium [component] experimenter interface [type] enhancement
|
Mapping nested objects (e.g., metrics to an experiment's metric assignments) is currently handled inconsistently throughout the code. @jessie-ross suggested we use [normalizr](https://github.com/paularmstrong/normalizr) for this.
See: https://github.com/Automattic/abacus/pull/198#discussion_r449975708
|
1.0
|
Apply consistent object normalisation across the codebase - Mapping nested objects (e.g., metrics to an experiment's metric assignments) is currently handled inconsistently throughout the code. @jessie-ross suggested we use [normalizr](https://github.com/paularmstrong/normalizr) for this.
See: https://github.com/Automattic/abacus/pull/198#discussion_r449975708
|
non_process
|
apply consistent object normalisation across the codebase mapping nested objects e g metrics to an experiment s metric assignments is currently handled inconsistently throughout the code jessie ross suggested we use for this see
| 0
|
3,666
| 6,694,824,296
|
IssuesEvent
|
2017-10-10 04:46:39
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
opened
|
View Watering Assignments - Finished Assignment Upload
|
enhancement process workflow
|
Create a data uploader to upload the finished assignment upload.
|
1.0
|
View Watering Assignments - Finished Assignment Upload - Create a data uploader to upload the finished assignment upload.
|
process
|
view watering assignments finished assignment upload create a data uploader to upload the finished assignment upload
| 1
|
57,769
| 14,219,807,762
|
IssuesEvent
|
2020-11-17 13:47:01
|
Automattic/wp-calypso
|
https://api.github.com/repos/Automattic/wp-calypso
|
closed
|
Modularise `userDevices`
|
Build
|
The `userDevices` portion of the state tree needs to be modularised. See the [modularised state documentation](https://github.com/Automattic/wp-calypso/blob/master/docs/modularized-state.md) for more details.
|
1.0
|
Modularise `userDevices` - The `userDevices` portion of the state tree needs to be modularised. See the [modularised state documentation](https://github.com/Automattic/wp-calypso/blob/master/docs/modularized-state.md) for more details.
|
non_process
|
modularise userdevices the userdevices portion of the state tree needs to be modularised see the for more details
| 0
|
13,338
| 15,800,886,211
|
IssuesEvent
|
2021-04-03 01:47:32
|
PyCQA/flake8
|
https://api.github.com/repos/PyCQA/flake8
|
closed
|
Order is nondeterministic with multiprocessing and piped output
|
bug:confirmed component:docs component:multiprocessing feature:accepted fix:committed fix:released
|
In GitLab by @quentinp on Nov 17, 2014, 03:24
Steps to reproduce:
* In a directory, create two files a.py and b.py with an error each (eg. both containing a=1).
* From the directory, run `flake8 -j2 . | tee` multiple times (``for i in `seq 1 10` ; do flake8 -j2 . | tee; echo; done`` can be useful).
Expected result:
The order is always the same, eg. `a.py`, then `b.py`. This is the case with `pep8` and `flake8 -j1`, or when the output is not piped. (By the way, I originally hit the error with `watch`, not `tee`).
Actual result:
The order is nondeterministic: sometimes `a.py` is first, sometimes `b.py` is first instead.
|
1.0
|
Order is nondeterministic with multiprocessing and piped output - In GitLab by @quentinp on Nov 17, 2014, 03:24
Steps to reproduce:
* In a directory, create two files a.py and b.py with an error each (eg. both containing a=1).
* From the directory, run `flake8 -j2 . | tee` multiple times (``for i in `seq 1 10` ; do flake8 -j2 . | tee; echo; done`` can be useful).
Expected result:
The order is always the same, eg. `a.py`, then `b.py`. This is the case with `pep8` and `flake8 -j1`, or when the output is not piped. (By the way, I originally hit the error with `watch`, not `tee`).
Actual result:
The order is nondeterministic: sometimes `a.py` is first, sometimes `b.py` is first instead.
|
process
|
order is nondeterministic with multiprocessing and piped output in gitlab by quentinp on nov steps to reproduce in a directory create two files a py and b py with an error each eg both containing a from the directory run tee multiple times for i in seq do tee echo done can be useful expected result the order is always the same eg a py then b py this is the case with and or when the output is not piped by the way i originally hit the error with watch not tee actual result the order is nondeterministic sometimes a py is first sometimes b py is first instead
| 1
|
42,271
| 9,199,573,387
|
IssuesEvent
|
2019-03-07 15:13:47
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
closed
|
Buttons contributed to the extensions toolbar should allow to specify "toggled" state
|
code-nav extensions feature-request
|
Many buttons contributed to the toolbar are to toggle something on or off, e.g. Coverage colors, soon code intel, ...
The on/off state can currently only be communicated through the label "Enable"/"Disable". This makes it impossible to omit the lable to save space, and looks very weird on GitHub where the buttons look like actual 3D buttons, that don't change into a "pressed" state when the setting is enabled.
There should be a field `"toggled"` in package.json for the button that is a boolean expression (and can read a setting for example) that makes the button toggled or not. On the Sourcegraph webapp, the state could be indicated by a background color change (something like `rgba(0, 0, 0, 0.5)` in light theme, `rgba(255, 255, 255, 0.5)` in dark theme).
I see this as a prerequisite to add an enable/disable for TypeScript because otherwise the experience is very weird.
Rel #1572
|
1.0
|
Buttons contributed to the extensions toolbar should allow to specify "toggled" state - Many buttons contributed to the toolbar are to toggle something on or off, e.g. Coverage colors, soon code intel, ...
The on/off state can currently only be communicated through the label "Enable"/"Disable". This makes it impossible to omit the lable to save space, and looks very weird on GitHub where the buttons look like actual 3D buttons, that don't change into a "pressed" state when the setting is enabled.
There should be a field `"toggled"` in package.json for the button that is a boolean expression (and can read a setting for example) that makes the button toggled or not. On the Sourcegraph webapp, the state could be indicated by a background color change (something like `rgba(0, 0, 0, 0.5)` in light theme, `rgba(255, 255, 255, 0.5)` in dark theme).
I see this as a prerequisite to add an enable/disable for TypeScript because otherwise the experience is very weird.
Rel #1572
|
non_process
|
buttons contributed to the extensions toolbar should allow to specify toggled state many buttons contributed to the toolbar are to toggle something on or off e g coverage colors soon code intel the on off state can currently only be communicated through the label enable disable this makes it impossible to omit the lable to save space and looks very weird on github where the buttons look like actual buttons that don t change into a pressed state when the setting is enabled there should be a field toggled in package json for the button that is a boolean expression and can read a setting for example that makes the button toggled or not on the sourcegraph webapp the state could be indicated by a background color change something like rgba in light theme rgba in dark theme i see this as a prerequisite to add an enable disable for typescript because otherwise the experience is very weird rel
| 0
|
13,241
| 15,708,282,161
|
IssuesEvent
|
2021-03-26 20:16:17
|
xatkit-bot-platform/xatkit-runtime
|
https://api.github.com/repos/xatkit-bot-platform/xatkit-runtime
|
opened
|
Remove @Ignore for language detection tests
|
Processors Testing
|
We have to ignore these tests for the moment because we don't have the infrastructure to run them in a CI/CD environment.
|
1.0
|
Remove @Ignore for language detection tests - We have to ignore these tests for the moment because we don't have the infrastructure to run them in a CI/CD environment.
|
process
|
remove ignore for language detection tests we have to ignore these tests for the moment because we don t have the infrastructure to run them in a ci cd environment
| 1
|
5,108
| 7,885,450,349
|
IssuesEvent
|
2018-06-27 12:31:08
|
Open-EO/openeo-api
|
https://api.github.com/repos/Open-EO/openeo-api
|
opened
|
Debugging and getting intermediate result and metadata
|
in discussion other process graphs processes
|
A big topic during the hackathon was debugging of our process graphs and getting information of intermediate result and other meta data. For example, after a filter one might want to know what data is left. GEE allows to add print() calls to get information about a variable with results between different steps. How could something like that be made available in openEO?
|
2.0
|
Debugging and getting intermediate result and metadata - A big topic during the hackathon was debugging of our process graphs and getting information of intermediate result and other meta data. For example, after a filter one might want to know what data is left. GEE allows to add print() calls to get information about a variable with results between different steps. How could something like that be made available in openEO?
|
process
|
debugging and getting intermediate result and metadata a big topic during the hackathon was debugging of our process graphs and getting information of intermediate result and other meta data for example after a filter one might want to know what data is left gee allows to add print calls to get information about a variable with results between different steps how could something like that be made available in openeo
| 1
|
288,556
| 24,915,737,836
|
IssuesEvent
|
2022-10-30 11:29:29
|
Jerrylum/ProtocolDiagram
|
https://api.github.com/repos/Jerrylum/ProtocolDiagram
|
closed
|
Generate diagram unit cases
|
test
|
I read the code on branch handle-diagram-generation. I see it is canvas-based. How are you going to unit test the code? Because I think comparing the canvas content in 2D in the test cases is not a good practice.
@JohnDoeAntler
|
1.0
|
Generate diagram unit cases - I read the code on branch handle-diagram-generation. I see it is canvas-based. How are you going to unit test the code? Because I think comparing the canvas content in 2D in the test cases is not a good practice.
@JohnDoeAntler
|
non_process
|
generate diagram unit cases i read the code on branch handle diagram generation i see it is canvas based how are you going to unit test the code because i think comparing the canvas content in in the test cases is not a good practice johndoeantler
| 0
|
500,247
| 14,494,248,168
|
IssuesEvent
|
2020-12-11 09:35:07
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
rule34.xxx - design is broken
|
browser-fenix engine-gecko priority-normal
|
<!-- @browser: Firefox Mobile 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/63472 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://rule34.xxx/index.php?page=post&s=list&tags=sinensian+
**Browser / Version**: Firefox Mobile 85.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206093054</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/8fc9e360-40b1-4a50-9f35-e288f86d8736)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
rule34.xxx - design is broken - <!-- @browser: Firefox Mobile 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/63472 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://rule34.xxx/index.php?page=post&s=list&tags=sinensian+
**Browser / Version**: Firefox Mobile 85.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206093054</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/8fc9e360-40b1-4a50-9f35-e288f86d8736)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
xxx design is broken url browser version firefox mobile operating system android tested another browser yes chrome problem type design is broken description images not loaded steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
93,754
| 3,910,681,267
|
IssuesEvent
|
2016-04-20 00:13:41
|
sdsu-its/fit-welcome
|
https://api.github.com/repos/sdsu-its/fit-welcome
|
closed
|
Cannot Swipe Card
|
Priority-Minor Type-Bug
|
When trying to login a user via a MagStripe Reader, the browser rejects the entry because it does not follow the specified format.
|
1.0
|
Cannot Swipe Card - When trying to login a user via a MagStripe Reader, the browser rejects the entry because it does not follow the specified format.
|
non_process
|
cannot swipe card when trying to login a user via a magstripe reader the browser rejects the entry because it does not follow the specified format
| 0
|
81,419
| 15,726,672,837
|
IssuesEvent
|
2021-03-29 11:39:04
|
GTNewHorizons/GT-New-Horizons-Modpack
|
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
|
closed
|
Evaluate multiblock ore drill speed/rebalance fluid use
|
Status: RFC (request for comment) Status: stale Type: Need Code changes
|
Multiblock ore drill speed is not very fast. Players would like to speed it up. Should it compare to Ender Quarry?
Also, would be interesting to have more fluid options.
So, suggestion:
~~- [ ] Reduce fluid use at the drill by 10x. Increase the lube/stone dust use in the mixer by 10x to compensate. This way players can use 256B drums to provide drilling fluid instead of needing bedrockium drums.~~ Players seem ok with this as-is.
- [ ] Increase drill base speed. For tier 2, base speed is 2s, adjust up/down appropriately for other tiers. Base speed - need to determine how this applies per tier. Should it all be respect to LV? Or LV = tier 1, MV = tier 2, etc?
- [ ] Add API to add custom advanced fluids, and how much faster they are compared to default. IE, fluid.advlube or whatever. Could be 4x faster, maybe add higher tier options.
- [ ] Let overclocking dictate base speed adjustments. Let advanced fluid speeds dictate how many blocks are picked per operation. Basic drilling fluid = 1x. Advanced = 4x. Maybe faster ones?
- [ ] Add a second fluid hatch for concrete? Let the machine place concrete in the drill holes. Dunno if that matters.
|
1.0
|
Evaluate multiblock ore drill speed/rebalance fluid use - Multiblock ore drill speed is not very fast. Players would like to speed it up. Should it compare to Ender Quarry?
Also, would be interesting to have more fluid options.
So, suggestion:
~~- [ ] Reduce fluid use at the drill by 10x. Increase the lube/stone dust use in the mixer by 10x to compensate. This way players can use 256B drums to provide drilling fluid instead of needing bedrockium drums.~~ Players seem ok with this as-is.
- [ ] Increase drill base speed. For tier 2, base speed is 2s, adjust up/down appropriately for other tiers. Base speed - need to determine how this applies per tier. Should it all be respect to LV? Or LV = tier 1, MV = tier 2, etc?
- [ ] Add API to add custom advanced fluids, and how much faster they are compared to default. IE, fluid.advlube or whatever. Could be 4x faster, maybe add higher tier options.
- [ ] Let overclocking dictate base speed adjustments. Let advanced fluid speeds dictate how many blocks are picked per operation. Basic drilling fluid = 1x. Advanced = 4x. Maybe faster ones?
- [ ] Add a second fluid hatch for concrete? Let the machine place concrete in the drill holes. Dunno if that matters.
|
non_process
|
evaluate multiblock ore drill speed rebalance fluid use multiblock ore drill speed is not very fast players would like to speed it up should it compare to ender quarry also would be interesting to have more fluid options so suggestion reduce fluid use at the drill by increase the lube stone dust use in the mixer by to compensate this way players can use drums to provide drilling fluid instead of needing bedrockium drums players seem ok with this as is increase drill base speed for tier base speed is adjust up down appropriately for other tiers base speed need to determine how this applies per tier should it all be respect to lv or lv tier mv tier etc add api to add custom advanced fluids and how much faster they are compared to default ie fluid advlube or whatever could be faster maybe add higher tier options let overclocking dictate base speed adjustments let advanced fluid speeds dictate how many blocks are picked per operation basic drilling fluid advanced maybe faster ones add a second fluid hatch for concrete let the machine place concrete in the drill holes dunno if that matters
| 0
|
9,287
| 13,100,884,561
|
IssuesEvent
|
2020-08-04 01:59:55
|
goharbor/harbor
|
https://api.github.com/repos/goharbor/harbor
|
closed
|
There are problems with the latest version pushing the same tag image and copy pull command
|
kind/requirement target/2.1.0 target/FC
|
**Expected behavior and actual behavior:**
1. How can harbor copy the pull command and get results using sha256 instead of tag.It really affects the experience.

2. I tag the same tag, which means the same version, but now the old version is still there, so there is a lot of garbage data

**Steps to reproduce the problem:**
1. When you click on the mirror pull command, you get sha256 instead of tag
2. Push the same tag image, check the mirror warehouse after success, and produce a lot of garbage data without tag
**Versions:**
Please specify the versions of following systems.
- harbor version: v2.0.0
- docker engine version: 19.03.8
- docker-compose version: 1.25.5
|
1.0
|
There are problems with the latest version pushing the same tag image and copy pull command - **Expected behavior and actual behavior:**
1. How can harbor copy the pull command and get results using sha256 instead of tag.It really affects the experience.

2. I tag the same tag, which means the same version, but now the old version is still there, so there is a lot of garbage data

**Steps to reproduce the problem:**
1. When you click on the mirror pull command, you get sha256 instead of tag
2. Push the same tag image, check the mirror warehouse after success, and produce a lot of garbage data without tag
**Versions:**
Please specify the versions of following systems.
- harbor version: v2.0.0
- docker engine version: 19.03.8
- docker-compose version: 1.25.5
|
non_process
|
there are problems with the latest version pushing the same tag image and copy pull command expected behavior and actual behavior how can harbor copy the pull command and get results using instead of tag it really affects the experience i tag the same tag which means the same version but now the old version is still there so there is a lot of garbage data steps to reproduce the problem when you click on the mirror pull command you get instead of tag push the same tag image check the mirror warehouse after success and produce a lot of garbage data without tag versions please specify the versions of following systems harbor version docker engine version docker compose version
| 0
|
48,675
| 13,392,365,480
|
IssuesEvent
|
2020-09-03 01:09:07
|
jgeraigery/shopping-list-react-pouchdb
|
https://api.github.com/repos/jgeraigery/shopping-list-react-pouchdb
|
opened
|
CVE-2020-7720 (High) detected in node-forge-0.9.0.tgz
|
security vulnerability
|
## CVE-2020-7720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/shopping-list-react-pouchdb/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/shopping-list-react-pouchdb/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.1.tgz (Root Library)
- webpack-dev-server-3.10.3.tgz
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:3.4.1;webpack-dev-server:3.10.3;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-7720","vulnerabilityDetails":"All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-7720 (High) detected in node-forge-0.9.0.tgz - ## CVE-2020-7720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/shopping-list-react-pouchdb/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/shopping-list-react-pouchdb/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.1.tgz (Root Library)
- webpack-dev-server-3.10.3.tgz
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function.
<p>Publish Date: 2020-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:3.4.1;webpack-dev-server:3.10.3;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-7720","vulnerabilityDetails":"All versions of package node-forge are vulnerable to Prototype Pollution via the util.setPath function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in node forge tgz cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file tmp ws scm shopping list react pouchdb package json path to vulnerable library tmp ws scm shopping list react pouchdb node modules node forge package json dependency hierarchy react scripts tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library vulnerability details all versions of package node forge are vulnerable to prototype pollution via the util setpath function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails all versions of package node forge are vulnerable to prototype pollution via the util setpath function vulnerabilityurl
| 0
|
177,548
| 14,631,909,130
|
IssuesEvent
|
2020-12-23 21:00:37
|
Josh194/FMCB
|
https://api.github.com/repos/Josh194/FMCB
|
closed
|
Decide on Subsystem General Structure.
|
documentation in progress
|
Create first draft for the general structure of the subsystem API. This should not include specific methods, structures, classes, etc, but should lay out the way in which subsystems will access the core buffer libraries, and be built (if applicable) with them.
|
1.0
|
Decide on Subsystem General Structure. - Create first draft for the general structure of the subsystem API. This should not include specific methods, structures, classes, etc, but should lay out the way in which subsystems will access the core buffer libraries, and be built (if applicable) with them.
|
non_process
|
decide on subsystem general structure create first draft for the general structure of the subsystem api this should not include specific methods structures classes etc but should lay out the way in which subsystems will access the core buffer libraries and be built if applicable with them
| 0
|
14,849
| 18,243,899,692
|
IssuesEvent
|
2021-10-01 15:52:05
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
filterprocessor: enable logs filtering by record level attributes
|
proc: filterprocessor
|
**Is your feature request related to a problem? Please describe.**
This is related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5237.
5237 tries to implement resource level log filtering with regexp whereas this issue aims to address filtering by record level attributes.
**Describe the solution you'd like**
~~I have a ready implementation that could be reviewed under https://github.com/pmalek-sumo/opentelemetry-collector-contrib/commit/11de2c03d3dbbde2e73b8a816318a2515c843985 (which would go on top of #5237 when it gets merged)~~
See: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/5279#issuecomment-925705513
**Describe alternatives you've considered**
There are no processors that would do it as far as I know.
|
1.0
|
filterprocessor: enable logs filtering by record level attributes - **Is your feature request related to a problem? Please describe.**
This is related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5237.
5237 tries to implement resource level log filtering with regexp whereas this issue aims to address filtering by record level attributes.
**Describe the solution you'd like**
~~I have a ready implementation that could be reviewed under https://github.com/pmalek-sumo/opentelemetry-collector-contrib/commit/11de2c03d3dbbde2e73b8a816318a2515c843985 (which would go on top of #5237 when it gets merged)~~
See: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/5279#issuecomment-925705513
**Describe alternatives you've considered**
There are no processors that would do it as far as I know.
|
process
|
filterprocessor enable logs filtering by record level attributes is your feature request related to a problem please describe this is related to tries to implement resource level log filtering with regexp whereas this issue aims to address filtering by record level attributes describe the solution you d like i have a ready implementation that could be reviewed under which would go on top of when it gets merged see describe alternatives you ve considered there are no processors that would do it as far as i know
| 1
|
9,758
| 12,741,968,970
|
IssuesEvent
|
2020-06-26 07:27:54
|
commudle/commudle-ng
|
https://api.github.com/repos/commudle/commudle-ng
|
closed
|
Simple online events
|
advanced api needed brainstorming needed enhancement process flow
|
Think and make the feature for simple online events. Right now Commudle supports both online and offline events, but it takes a lot of details, including forms to set it up.
Let's to make a simple plug and play event feature (the current event setup would be the advanced version of it).
|
1.0
|
Simple online events - Think and make the feature for simple online events. Right now Commudle supports both online and offline events, but it takes a lot of details, including forms to set it up.
Let's to make a simple plug and play event feature (the current event setup would be the advanced version of it).
|
process
|
simple online events think and make the feature for simple online events right now commudle supports both online and offline events but it takes a lot of details including forms to set it up let s to make a simple plug and play event feature the current event setup would be the advanced version of it
| 1
|
12,879
| 15,268,816,220
|
IssuesEvent
|
2021-02-22 11:55:48
|
threefoldfoundation/home
|
https://api.github.com/repos/threefoldfoundation/home
|
closed
|
Marketplace Web - cleanup
|
process_duplicate
|
- [ ] now.threefold.io : which is the commercial website about marketplace and has all solutions described (team promotion)
- [ ] content from gridsome wiki ex now.threefold.io -> the new now.threefold.io which is based on new web framework (team promotion)
- [ ] on each commercial page there is "deploy now button" -> go to auth page, see who user is, redirect to their evdc and their marketplace for that solution
- [ ] if you don't exist, ask to install TF Connect
|
1.0
|
Marketplace Web - cleanup - - [ ] now.threefold.io : which is the commercial website about marketplace and has all solutions described (team promotion)
- [ ] content from gridsome wiki ex now.threefold.io -> the new now.threefold.io which is based on new web framework (team promotion)
- [ ] on each commercial page there is "deploy now button" -> go to auth page, see who user is, redirect to their evdc and their marketplace for that solution
- [ ] if you don't exist, ask to install TF Connect
|
process
|
marketplace web cleanup now threefold io which is the commercial website about marketplace and has all solutions described team promotion content from gridsome wiki ex now threefold io the new now threefold io which is based on new web framework team promotion on each commercial page there is deploy now button go to auth page see who user is redirect to their evdc and their marketplace for that solution if you don t exist ask to install tf connect
| 1
|
27,879
| 6,911,094,577
|
IssuesEvent
|
2017-11-28 06:34:29
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
Joomla 4 save & close two click procedure.
|
No Code Attached Yet
|
### Steps to reproduce the issue
open module and save
### Expected result
save and close being the default / quick action
### Actual result
save is the quick action
### System information (as much as possible)
Joomla! 4.0.0-dev Development [ Amani ] 19-November-2017 16:34 GMT
### Additional comments
It's completely counterintuitive & counter productive to have to hit a dropdown to save & close.
it's far more common that people make all their changes and then save, then it is to make incremental saves. These button choices should be reversed!
|
1.0
|
Joomla 4 save & close two click procedure. - ### Steps to reproduce the issue
open module and save
### Expected result
save and close being the default / quick action
### Actual result
save is the quick action
### System information (as much as possible)
Joomla! 4.0.0-dev Development [ Amani ] 19-November-2017 16:34 GMT
### Additional comments
It's completely counterintuitive & counter productive to have to hit a dropdown to save & close.
it's far more common that people make all their changes and then save, then it is to make incremental saves. These button choices should be reversed!
|
non_process
|
joomla save close two click procedure steps to reproduce the issue open module and save expected result save and close being the default quick action actual result save is the quick action system information as much as possible joomla dev development november gmt additional comments it s completely counterintuitive counter productive to have to hit a dropdown to save close it s far more common that people make all their changes and then save then it is to make incremental saves these button choices should be reversed
| 0
|
5,191
| 7,973,705,787
|
IssuesEvent
|
2018-07-17 00:45:54
|
bugmark/exchange
|
https://api.github.com/repos/bugmark/exchange
|
closed
|
Use of branches in repo?
|
9 question DEV PROCESS
|
It appears that the `master` branch is ahead of the `dev` branch. This is not intuitive for me and I would like clarity of how branches are used in this repo?
I ask, because the pull request #82 is against the dev branch, but people will start with the master branch, and if there is no path from dev to master, then I'm confused. Also, if the dev branch is meant for development environment, why should it not have the latest version of features?
|
1.0
|
Use of branches in repo? - It appears that the `master` branch is ahead of the `dev` branch. This is not intuitive for me and I would like clarity of how branches are used in this repo?
I ask, because the pull request #82 is against the dev branch, but people will start with the master branch, and if there is no path from dev to master, then I'm confused. Also, if the dev branch is meant for development environment, why should it not have the latest version of features?
|
process
|
use of branches in repo it appears that the master branch is ahead of the dev branch this is not intuitive for me and i would like clarity of how branches are used in this repo i ask because the pull request is against the dev branch but people will start with the master branch and if there is no path from dev to master then i m confused also if the dev branch is meant for development environment why should it not have the latest version of features
| 1
|
8,710
| 2,611,536,919
|
IssuesEvent
|
2015-02-27 06:06:27
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
"The best shot award was won by <b>%1</b> with <b>%2</b> pts." needs plural forms
|
auto-migrated Priority-Medium Type-Defect
|
```
"The best shot award was won by <b>%1</b> with <b>%2</b> pts." in PageGameStats
needs marking for plural forms, like you already did with "<b>%1</b> thought
it's good to shoot his own hedgehogs with <b>%2</b> pts."
```
Original issue reported on code.google.com by `f...@foramnagaidhlig.net` on 4 Aug 2014 at 1:19
|
1.0
|
"The best shot award was won by <b>%1</b> with <b>%2</b> pts." needs plural forms - ```
"The best shot award was won by <b>%1</b> with <b>%2</b> pts." in PageGameStats
needs marking for plural forms, like you already did with "<b>%1</b> thought
it's good to shoot his own hedgehogs with <b>%2</b> pts."
```
Original issue reported on code.google.com by `f...@foramnagaidhlig.net` on 4 Aug 2014 at 1:19
|
non_process
|
the best shot award was won by with pts needs plural forms the best shot award was won by with pts in pagegamestats needs marking for plural forms like you already did with thought it s good to shoot his own hedgehogs with pts original issue reported on code google com by f foramnagaidhlig net on aug at
| 0
|
2,801
| 5,731,271,047
|
IssuesEvent
|
2017-04-21 11:54:22
|
g8os/core0
|
https://api.github.com/repos/g8os/core0
|
closed
|
container.bash depends on the $PATH environment variable
|
process_wontfix type_bug
|
branch 1.1.0-alpha
```python
In [47]: cl.container.create("https://hub.gig.tech/gig-official-apps/ardb-rocksdb.flist", mount={'/mnt/dev-sda1':'/mnt'}, host_network=True)
Out[47]: 1
In [50]: container = cl.container.client(1)
In [53]: container.bash("ip a").get()
Out[53]:
STATE: ERROR
STDOUT:
STDERR:
DATA:
exec: "sh": executable file not found in $PATH
```
Wouldn't it better to actually use `/bin/bash` ?
|
1.0
|
container.bash depends on the $PATH environment variable - branch 1.1.0-alpha
```python
In [47]: cl.container.create("https://hub.gig.tech/gig-official-apps/ardb-rocksdb.flist", mount={'/mnt/dev-sda1':'/mnt'}, host_network=True)
Out[47]: 1
In [50]: container = cl.container.client(1)
In [53]: container.bash("ip a").get()
Out[53]:
STATE: ERROR
STDOUT:
STDERR:
DATA:
exec: "sh": executable file not found in $PATH
```
Wouldn't it better to actually use `/bin/bash` ?
|
process
|
container bash depends on the path environment variable branch alpha python in cl container create mount mnt dev mnt host network true out in container cl container client in container bash ip a get out state error stdout stderr data exec sh executable file not found in path wouldn t it better to actually use bin bash
| 1
|
3,803
| 6,782,426,395
|
IssuesEvent
|
2017-10-30 08:02:37
|
decidim/decidim
|
https://api.github.com/repos/decidim/decidim
|
closed
|
Improve process metadata field naming and ordering
|
component: processes good first issue hacktoberfest team: i18n
|
# This is a Feature Proposal
#### :tophat: Description
* Current name for the metadata that is displayed for participatory processes is very confusing.
* Example from http://staging.decidim.codegram.com/processes/omnis-sint :

* What needs to be done is the following:
- [ ] change name Participatory scope --> What is decided
- [ ] change name Participatory structure --> How is it decided
- [ ] change name Target --> Who participates
- [ ] local area --> Organization Area
- [ ] change order to have it as follows: 1. What is decided, 2. Who participates, 3. How is it decided, 4. Ámbit, 5. Organization Area, 6. Developer Group, 7. Start Date, 8. End date
- [ ] Mention @xabier to update the translations :)
|
1.0
|
Improve process metadata field naming and ordering - # This is a Feature Proposal
#### :tophat: Description
* Current name for the metadata that is displayed for participatory processes is very confusing.
* Example from http://staging.decidim.codegram.com/processes/omnis-sint :

* What needs to be done is the following:
- [ ] change name Participatory scope --> What is decided
- [ ] change name Participatory structure --> How is it decided
- [ ] change name Target --> Who participates
- [ ] local area --> Organization Area
- [ ] change order to have it as follows: 1. What is decided, 2. Who participates, 3. How is it decided, 4. Ámbit, 5. Organization Area, 6. Developer Group, 7. Start Date, 8. End date
- [ ] Mention @xabier to update the translations :)
|
process
|
improve process metadata field naming and ordering this is a feature proposal tophat description current name for the metadata that is displayed for participatory processes is very confusing example from what needs to be done is the following change name participatory scope what is decided change name participatory structure how is it decided change name target who participates local area organization area change order to have it as follows what is decided who participates how is it decided ámbit organization area developer group start date end date mention xabier to update the translations
| 1
|
12,532
| 14,972,344,762
|
IssuesEvent
|
2021-01-27 22:41:01
|
BootBlock/FileSieve
|
https://api.github.com/repos/BootBlock/FileSieve
|
opened
|
Add a transaction log of all I/O changes made after processing
|
backend-core enhancement processing
|
If a detailed-enough log is kept of a full “processing”, then maybe that’ll allow the possibility of “undoing” a processing after-the-fact?
Would need proper planning and checks in place as it could get somewhat messy if anything unexpected happens.
|
1.0
|
Add a transaction log of all I/O changes made after processing - If a detailed-enough log is kept of a full “processing”, then maybe that’ll allow the possibility of “undoing” a processing after-the-fact?
Would need proper planning and checks in place as it could get somewhat messy if anything unexpected happens.
|
process
|
add a transaction log of all i o changes made after processing if a detailed enough log is kept of a full “processing” then maybe that’ll allow the possibility of “undoing” a processing after the fact would need proper planning and checks in place as it could get somewhat messy if anything unexpected happens
| 1
|
512,488
| 14,897,962,395
|
IssuesEvent
|
2021-01-21 12:30:02
|
kubermatic/kubermatic
|
https://api.github.com/repos/kubermatic/kubermatic
|
closed
|
Validation: I can create Azure cluster without specifying node size
|
kind/bug priority/normal sig/ui
|
I post an azure cluster without specifying node size. Expected that API would validate required fields.
|
1.0
|
Validation: I can create Azure cluster without specifying node size - I post an azure cluster without specifying node size. Expected that API would validate required fields.
|
non_process
|
validation i can create azure cluster without specifying node size i post an azure cluster without specifying node size expected that api would validate required fields
| 0
|
251,225
| 27,151,001,180
|
IssuesEvent
|
2023-02-17 01:14:39
|
DeepikaKaranji/ContainerApplication-Security-With-JenkinsCI
|
https://api.github.com/repos/DeepikaKaranji/ContainerApplication-Security-With-JenkinsCI
|
opened
|
CVE-2023-23934 (Low) detected in Werkzeug-1.0.1-py2.py3-none-any.whl
|
security vulnerability
|
## CVE-2023-23934 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary>
<p>The comprehensive WSGI web application library.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p>
<p>Path to vulnerable library: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p>
<p>
Dependency Hierarchy:
- Flask-1.1.2-py2.py3-none-any.whl (Root Library)
- :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Werkzeug is a comprehensive WSGI web application library. Browsers may allow "nameless" cookies that look like `=value` instead of `key=value`. A vulnerable browser may allow a compromised application on an adjacent subdomain to exploit this to set a cookie like `=__Host-test=bad` for another subdomain. Werkzeug prior to 2.2.3 will parse the cookie `=__Host-test=bad` as __Host-test=bad`. If a Werkzeug application is running next to a vulnerable or malicious subdomain which sets such a cookie using a vulnerable browser, the Werkzeug application will see the bad cookie value but the valid cookie key. The issue is fixed in Werkzeug 2.2.3.
<p>Publish Date: 2023-02-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23934>CVE-2023-23934</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23934">https://www.cve.org/CVERecord?id=CVE-2023-23934</a></p>
<p>Release Date: 2023-02-14</p>
<p>Fix Resolution: Werkzeug - 2.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-23934 (Low) detected in Werkzeug-1.0.1-py2.py3-none-any.whl - ## CVE-2023-23934 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary>
<p>The comprehensive WSGI web application library.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p>
<p>Path to vulnerable library: /tmp/ws-scm/Container-Security-With-JenkinsCI/requirements.txt</p>
<p>
Dependency Hierarchy:
- Flask-1.1.2-py2.py3-none-any.whl (Root Library)
- :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Werkzeug is a comprehensive WSGI web application library. Browsers may allow "nameless" cookies that look like `=value` instead of `key=value`. A vulnerable browser may allow a compromised application on an adjacent subdomain to exploit this to set a cookie like `=__Host-test=bad` for another subdomain. Werkzeug prior to 2.2.3 will parse the cookie `=__Host-test=bad` as __Host-test=bad`. If a Werkzeug application is running next to a vulnerable or malicious subdomain which sets such a cookie using a vulnerable browser, the Werkzeug application will see the bad cookie value but the valid cookie key. The issue is fixed in Werkzeug 2.2.3.
<p>Publish Date: 2023-02-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23934>CVE-2023-23934</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23934">https://www.cve.org/CVERecord?id=CVE-2023-23934</a></p>
<p>Release Date: 2023-02-14</p>
<p>Fix Resolution: Werkzeug - 2.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve low detected in werkzeug none any whl cve low severity vulnerability vulnerable library werkzeug none any whl the comprehensive wsgi web application library library home page a href path to dependency file tmp ws scm container security with jenkinsci requirements txt path to vulnerable library tmp ws scm container security with jenkinsci requirements txt dependency hierarchy flask none any whl root library x werkzeug none any whl vulnerable library vulnerability details werkzeug is a comprehensive wsgi web application library browsers may allow nameless cookies that look like value instead of key value a vulnerable browser may allow a compromised application on an adjacent subdomain to exploit this to set a cookie like host test bad for another subdomain werkzeug prior to will parse the cookie host test bad as host test bad if a werkzeug application is running next to a vulnerable or malicious subdomain which sets such a cookie using a vulnerable browser the werkzeug application will see the bad cookie value but the valid cookie key the issue is fixed in werkzeug publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution werkzeug step up your open source security game with mend
| 0
|
520,865
| 15,096,051,356
|
IssuesEvent
|
2021-02-07 13:37:59
|
RocketMeet/RocketMeet-client
|
https://api.github.com/repos/RocketMeet/RocketMeet-client
|
reopened
|
Comma-separation handler on poll pages not working on mobile
|
Priority: High bug
|
From reports by different users, it looks like it could be a browser-related issue too.
|
1.0
|
Comma-separation handler on poll pages not working on mobile - From reports by different users, it looks like it could be a browser-related issue too.
|
non_process
|
comma separation handler on poll pages not working on mobile from reports by different users it looks like it could be a browser related issue too
| 0
|
5,701
| 8,563,612,108
|
IssuesEvent
|
2018-11-09 14:33:26
|
easy-software-ufal/annotations_repos
|
https://api.github.com/repos/easy-software-ufal/annotations_repos
|
opened
|
aspnet/JsonPatch JsonPatchDocument.Replace() yields invalid path when [JsonProperty] is used (1.1.0)
|
C# RMP test wrong processing
|
Issue: `https://github.com/aspnet/JsonPatch/issues/50`
PR: `https://github.com/aspnet/JsonPatch/commit/e8452821b9a7689853e36c9a10ad4b8c6281817c`
|
1.0
|
aspnet/JsonPatch JsonPatchDocument.Replace() yields invalid path when [JsonProperty] is used (1.1.0) - Issue: `https://github.com/aspnet/JsonPatch/issues/50`
PR: `https://github.com/aspnet/JsonPatch/commit/e8452821b9a7689853e36c9a10ad4b8c6281817c`
|
process
|
aspnet jsonpatch jsonpatchdocument replace yields invalid path when is used issue pr
| 1
|
7,334
| 10,469,054,019
|
IssuesEvent
|
2019-09-22 18:00:02
|
produvia/ai-platform
|
https://api.github.com/repos/produvia/ai-platform
|
closed
|
Machine Translation
|
natural-language-processing task wontfix
|
# Goal(s)
- Translate a sentence in a source language to a different target language
# Input(s)
- Sentence
# Output(s)
- Sentence
# Objective Function(s)
- TBD
|
1.0
|
Machine Translation - # Goal(s)
- Translate a sentence in a source language to a different target language
# Input(s)
- Sentence
# Output(s)
- Sentence
# Objective Function(s)
- TBD
|
process
|
machine translation goal s translate a sentence in a source language to a different target language input s sentence output s sentence objective function s tbd
| 1
|
209
| 2,636,660,179
|
IssuesEvent
|
2015-03-10 06:39:31
|
fabric8io/fabric8
|
https://api.github.com/repos/fabric8io/fabric8
|
closed
|
Auto-detect base package of deployed Spring Boot app
|
java container micro services process manager V1
|
Now we need to specify the package or base class of the deployed Spring Boot application. For example:
profile-edit --pid=io.fabric8.container.java/jvmArguments=-Dspring.main.sources=io.fabric8.process.spring.boot.itests.service.invoicing invoicing
I don't like it, because this is error-prone part of the application setup (typo in package name, etc). It will be nice if Fabric8 could try to autodetect base package of the installed jar. This is do-able with some basic jar file scanning.
For example:
com
example
project
MyClass1.class
MyClass2.class
Should return `com.example.project`.
|
1.0
|
Auto-detect base package of deployed Spring Boot app - Now we need to specify the package or base class of the deployed Spring Boot application. For example:
profile-edit --pid=io.fabric8.container.java/jvmArguments=-Dspring.main.sources=io.fabric8.process.spring.boot.itests.service.invoicing invoicing
I don't like it, because this is error-prone part of the application setup (typo in package name, etc). It will be nice if Fabric8 could try to autodetect base package of the installed jar. This is do-able with some basic jar file scanning.
For example:
com
example
project
MyClass1.class
MyClass2.class
Should return `com.example.project`.
|
process
|
auto detect base package of deployed spring boot app now we need to specify the package or base class of the deployed spring boot application for example profile edit pid io container java jvmarguments dspring main sources io process spring boot itests service invoicing invoicing i don t like it because this is error prone part of the application setup typo in package name etc it will be nice if could try to autodetect base package of the installed jar this is do able with some basic jar file scanning for example com example project class class should return com example project
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.