Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
167,036
| 14,099,697,038
|
IssuesEvent
|
2020-11-06 02:06:03
|
matplotlib/matplotlib
|
https://api.github.com/repos/matplotlib/matplotlib
|
opened
|
Current `tight_layout` example raises UserWarning
|
API Changes Documentation
|
<!--To help us understand and resolve your issue, please fill out the form to the best of your ability.-->
<!--You can feel free to delete the sections that do not apply.-->
### Problem
As can be seen in the [tight layout docs](https://matplotlib.org/tutorials/intermediate/tight_layout_guide.html) (Ctrl-F for "Out:"), `GridSpec.tight_layout` raises a `UserWarning` whenever there are `Axes` in a figure that come from other `GridSpec` besides `self`.
Is this the intended API?
If so, is the use case being demonstrated when that warning is raised (using different subplot specs to render in different `rect`'s within a single figure, without needed nested gridspec's) deprecated? Should we be instead encouraging users to use nested GridSpec's instead for this use-case?
If not, then should there be a flag to `tight_layout` that tells it to ignore other `Axes` in the figure, like a "I know what I'm doing" flag? I can't see a way for `tight_layout.get_subplotspec_list` to differentiate, currently, (and this may be due to a lack of my own imagination) between the two relevant options:
1. the user accidentally passed in a subplotspec that doesn't encompass the entire figure (I guess that's what the warning is for), OR
2. whether they intentionally are passing in gridspec that only describes a subset of Axes (as it's being used in the example)
|
1.0
|
Current `tight_layout` example raises UserWarning - <!--To help us understand and resolve your issue, please fill out the form to the best of your ability.-->
<!--You can feel free to delete the sections that do not apply.-->
### Problem
As can be seen in the [tight layout docs](https://matplotlib.org/tutorials/intermediate/tight_layout_guide.html) (Ctrl-F for "Out:"), `GridSpec.tight_layout` raises a `UserWarning` whenever there are `Axes` in a figure that come from other `GridSpec` besides `self`.
Is this the intended API?
If so, is the use case being demonstrated when that warning is raised (using different subplot specs to render in different `rect`'s within a single figure, without needed nested gridspec's) deprecated? Should we be instead encouraging users to use nested GridSpec's instead for this use-case?
If not, then should there be a flag to `tight_layout` that tells it to ignore other `Axes` in the figure, like a "I know what I'm doing" flag? I can't see a way for `tight_layout.get_subplotspec_list` to differentiate, currently, (and this may be due to a lack of my own imagination) between the two relevant options:
1. the user accidentally passed in a subplotspec that doesn't encompass the entire figure (I guess that's what the warning is for), OR
2. whether they intentionally are passing in gridspec that only describes a subset of Axes (as it's being used in the example)
|
non_process
|
current tight layout example raises userwarning problem as can be seen in the ctrl f for out gridspec tight layout raises a userwarning whenever there are axes in a figure that come from other gridspec besides self is this the intended api if so is the use case being demonstrated when that warning is raised using different subplot specs to render in different rect s within a single figure without needed nested gridspec s deprecated should we be instead encouraging users to use nested gridspec s instead for this use case if not then should there be a flag to tight layout that tells it to ignore other axes in the figure like a i know what i m doing flag i can t see a way for tight layout get subplotspec list to differentiate currently and this may be due to a lack of my own imagination between the two relevant options the user accidentally passed in a subplotspec that doesn t encompass the entire figure i guess that s what the warning is for or whether they intentionally are passing in gridspec that only describes a subset of axes as it s being used in the example
| 0
|
39,926
| 10,421,705,655
|
IssuesEvent
|
2019-09-16 07:06:03
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Fix Javadoc warnings
|
C: Build C: Documentation E: All Editions P: Low T: Defect
|
The build currently complains about the following Javadoc warnings:
```
[WARNING] Javadoc Warnings
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:60: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:61: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:63: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:69: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:87: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:88: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:90: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:95: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:96: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Block.java:67: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Configuration.java:275: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DAO.java:57: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DataType.java:346: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:66: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:67: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag >
```
|
1.0
|
Fix Javadoc warnings - The build currently complains about the following Javadoc warnings:
```
[WARNING] Javadoc Warnings
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:60: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:61: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:63: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:69: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:87: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:88: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:90: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:95: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:96: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Block.java:67: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Configuration.java:275: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DAO.java:57: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DataType.java:346: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:66: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:67: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag >
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag <
[WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag >
```
|
non_process
|
fix javadoc warnings the build currently complains about the following javadoc warnings javadoc warnings c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq block java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq commontableexpression java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq commontableexpression java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq configuration java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converter java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converter java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dao java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq datatype java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq declaration java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq declaration java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq derivedcolumnlist java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq derivedcolumnlist java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag
| 0
|
9,109
| 12,192,306,098
|
IssuesEvent
|
2020-04-29 12:44:07
|
naoki-shigehisa/paper
|
https://api.github.com/repos/naoki-shigehisa/paper
|
opened
|
Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems
|
2019 Gaussian Process recommendation
|
## 0. 論文
タイトル:[Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems](https://arxiv.org/abs/1912.09593)
著者:

arXiv投稿日:2019/12/19
学会/ジャーナル:
## 1. どんなもの?
## 2. 先行研究と比べてどこがすごい?
## 3. 技術や手法のキモはどこ?
## 4. どうやって有効だと検証した?
## 5. 議論はある?
## 6. 次に読むべき論文は?
|
1.0
|
Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems - ## 0. 論文
タイトル:[Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems](https://arxiv.org/abs/1912.09593)
著者:

arXiv投稿日:2019/12/19
学会/ジャーナル:
## 1. どんなもの?
## 2. 先行研究と比べてどこがすごい?
## 3. 技術や手法のキモはどこ?
## 4. どうやって有効だと検証した?
## 5. 議論はある?
## 6. 次に読むべき論文は?
|
process
|
gaussian process latent variable model factorization for context aware recommender systems 論文 タイトル: 著者: arxiv投稿日: 学会 ジャーナル: どんなもの? 先行研究と比べてどこがすごい? 技術や手法のキモはどこ? どうやって有効だと検証した? 議論はある? 次に読むべき論文は?
| 1
|
20,243
| 26,861,732,740
|
IssuesEvent
|
2023-02-03 19:03:22
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
reopened
|
multiprocessing's default posix start method of `'fork'` is broken: change to `'spawn'`
|
type-feature 3.12 expert-multiprocessing
|
BPO | [40379](https://bugs.python.org/issue40379)
--- | :---
Nosy | @pitrou, @mgorny, @Julian, @wimglenn, @applio, @itamarst
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2020-04-24.18:22:23.389>
labels = ['3.8', 'type-bug', '3.7', '3.9']
title = "multiprocessing's default start method of fork()-without-exec() is broken"
updated_at = <Date 2022-02-11.16:13:53.872>
user = 'https://bugs.python.org/itamarst'
```
bugs.python.org fields:
```python
activity = <Date 2022-02-11.16:13:53.872>
actor = 'mgorny'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = []
creation = <Date 2020-04-24.18:22:23.389>
creator = 'itamarst'
dependencies = []
files = []
hgrepos = []
issue_num = 40379
keywords = []
message_count = 11.0
messages = ['367210', '367211', '368173', '380478', '392358', '392501', '392503', '392506', '392507', '392508', '413081']
nosy_count = 8.0
nosy_names = ['pitrou', 'mgorny', 'Julian', 'wim.glenn', 'itamarst', 'davin', 'itamarst2', 'aduncan']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'behavior'
url = 'https://bugs.python.org/issue40379'
versions = ['Python 3.5', 'Python 3.6', 'Python 3.7', 'Python 3.8', 'Python 3.9']
```
</p></details>
<!-- gh-linked-prs -->
### Linked PRs
* gh-100618
* gh-101530
<!-- /gh-linked-prs -->
|
1.0
|
multiprocessing's default posix start method of `'fork'` is broken: change to `'spawn'` - BPO | [40379](https://bugs.python.org/issue40379)
--- | :---
Nosy | @pitrou, @mgorny, @Julian, @wimglenn, @applio, @itamarst
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = None
closed_at = None
created_at = <Date 2020-04-24.18:22:23.389>
labels = ['3.8', 'type-bug', '3.7', '3.9']
title = "multiprocessing's default start method of fork()-without-exec() is broken"
updated_at = <Date 2022-02-11.16:13:53.872>
user = 'https://bugs.python.org/itamarst'
```
bugs.python.org fields:
```python
activity = <Date 2022-02-11.16:13:53.872>
actor = 'mgorny'
assignee = 'none'
closed = False
closed_date = None
closer = None
components = []
creation = <Date 2020-04-24.18:22:23.389>
creator = 'itamarst'
dependencies = []
files = []
hgrepos = []
issue_num = 40379
keywords = []
message_count = 11.0
messages = ['367210', '367211', '368173', '380478', '392358', '392501', '392503', '392506', '392507', '392508', '413081']
nosy_count = 8.0
nosy_names = ['pitrou', 'mgorny', 'Julian', 'wim.glenn', 'itamarst', 'davin', 'itamarst2', 'aduncan']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = 'behavior'
url = 'https://bugs.python.org/issue40379'
versions = ['Python 3.5', 'Python 3.6', 'Python 3.7', 'Python 3.8', 'Python 3.9']
```
</p></details>
<!-- gh-linked-prs -->
### Linked PRs
* gh-100618
* gh-101530
<!-- /gh-linked-prs -->
|
process
|
multiprocessing s default posix start method of fork is broken change to spawn bpo nosy pitrou mgorny julian wimglenn applio itamarst note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title multiprocessing s default start method of fork without exec is broken updated at user bugs python org fields python activity actor mgorny assignee none closed false closed date none closer none components creation creator itamarst dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type behavior url versions linked prs gh gh
| 1
|
497,319
| 14,367,968,311
|
IssuesEvent
|
2020-12-01 07:38:16
|
teamforus/forus
|
https://api.github.com/repos/teamforus/forus
|
closed
|
CR: Webshop - remove "Hoe het werkt" below header description
|
Priority: Must have
|
@maxvisser commented on [Tue Sep 15 2020](https://github.com/teamforus/general/issues/445)
Learn more about change requests here: https://bit.ly/39CWeEE
### Requested by:
- Geertruidenberg
### Change description
I don't think such a button really fits there if there is a menu at the top of the page.
The button 'how it works' and 'explanation' in the menu link to the same page.
I would delete "Hoe werkt het".

|
1.0
|
CR: Webshop - remove "Hoe het werkt" below header description - @maxvisser commented on [Tue Sep 15 2020](https://github.com/teamforus/general/issues/445)
Learn more about change requests here: https://bit.ly/39CWeEE
### Requested by:
- Geertruidenberg
### Change description
I don't think such a button really fits there if there is a menu at the top of the page.
The button 'how it works' and 'explanation' in the menu link to the same page.
I would delete "Hoe werkt het".

|
non_process
|
cr webshop remove hoe het werkt below header description maxvisser commented on learn more about change requests here requested by geertruidenberg change description i don t think such a button really fits there if there is a menu at the top of the page the button how it works and explanation in the menu link to the same page i would delete hoe werkt het
| 0
|
22,568
| 31,790,020,810
|
IssuesEvent
|
2023-09-13 02:00:10
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Wed, 13 Sep 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### SoccerNet 2023 Challenges Results
- **Authors:** Anthony Cioppa, Silvio Giancola, Vladimir Somers, Floriane Magera, Xin Zhou, Hassan Mkhallati, Adrien Deliège, Jan Held, Carlos Hinojosa, Amir M. Mansourian, Pierre Miralles, Olivier Barnich, Christophe De Vleeschouwer, Alexandre Alahi, Bernard Ghanem, Marc Van Droogenbroeck, Abdullah Kamal, Adrien Maglo, Albert Clapés, Amr Abdelaziz, Artur Xarles, Astrid Orcesi, Atom Scott, Bin Liu, Byoungkwon Lim, Chen Chen, Fabian Deuser, Feng Yan, Fufu Yu, Gal Shitrit, Guanshuo Wang, Gyusik Choi, Hankyul Kim, Hao Guo, Hasby Fahrudin, Hidenari Koguchi, Håkan Ardö, Ibrahim Salah, Ido Yerushalmy, Iftikar Muhammad, Ikuma Uchida, Ishay Be'ery, Jaonary Rabarisoa, Jeongae Lee, Jiajun Fu, Jianqin Yin, Jinghang Xu, Jongho Nang, Julien Denize, Junjie Li, Junpei Zhang, Juntae Kim, Kamil Synowiec, et al. (49 additional authors not shown)
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.06006
- **Pdf link:** https://arxiv.org/pdf/2309.06006
- **Abstract**
The SoccerNet 2023 challenges were the third annual video understanding challenges organized by the SoccerNet team. For this third edition, the challenges were composed of seven vision-based tasks split into three main themes. The first theme, broadcast video understanding, is composed of three high-level tasks related to describing events occurring in the video broadcasts: (1) action spotting, focusing on retrieving all timestamps related to global actions in soccer, (2) ball action spotting, focusing on retrieving all timestamps related to the soccer ball change of state, and (3) dense video captioning, focusing on describing the broadcast with natural language and anchored timestamps. The second theme, field understanding, relates to the single task of (4) camera calibration, focusing on retrieving the intrinsic and extrinsic camera parameters from images. The third and last theme, player understanding, is composed of three low-level tasks related to extracting information about the players: (5) re-identification, focusing on retrieving the same players across multiple views, (6) multiple object tracking, focusing on tracking players and the ball through unedited video streams, and (7) jersey number recognition, focusing on recognizing the jersey number of players from tracklets. Compared to the previous editions of the SoccerNet challenges, tasks (2-3-7) are novel, including new annotations and data, task (4) was enhanced with more data and annotations, and task (6) now focuses on end-to-end approaches. More information on the tasks, challenges, and leaderboards are available on https://www.soccer-net.org. Baselines and development kits can be found on https://github.com/SoccerNet.
### JOADAA: joint online action detection and action anticipation
- **Authors:** Mohammed Guermal, Francois Bremond, Rui Dai, Abid Ali
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.06130
- **Pdf link:** https://arxiv.org/pdf/2309.06130
- **Abstract**
Action anticipation involves forecasting future actions by connecting past events to future ones. However, this reasoning ignores the real-life hierarchy of events which is considered to be composed of three main parts: past, present, and future. We argue that considering these three main parts and their dependencies could improve performance. On the other hand, online action detection is the task of predicting actions in a streaming manner. In this case, one has access only to the past and present information. Therefore, in online action detection (OAD) the existing approaches miss semantics or future information which limits their performance. To sum up, for both of these tasks, the complete set of knowledge (past-present-future) is missing, which makes it challenging to infer action dependencies, therefore having low performances. To address this limitation, we propose to fuse both tasks into a single uniform architecture. By combining action anticipation and online action detection, our approach can cover the missing dependencies of future information in online action detection. This method referred to as JOADAA, presents a uniform model that jointly performs action anticipation and online action detection. We validate our proposed model on three challenging datasets: THUMOS'14, which is a sparsely annotated dataset with one action per time step, CHARADES, and Multi-THUMOS, two densely annotated datasets with more complex scenarios. JOADAA achieves SOTA results on these benchmarks for both tasks.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Rice Plant Disease Detection and Diagnosis using Deep Convolutional Neural Networks and Multispectral Imaging
- **Authors:** Yara Ali Alnaggar, Ahmad Sebaq, Karim Amer, ElSayed Naeem, Mohamed Elhelw
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2309.05818
- **Pdf link:** https://arxiv.org/pdf/2309.05818
- **Abstract**
Rice is considered a strategic crop in Egypt as it is regularly consumed in the Egyptian people's diet. Even though Egypt is the highest rice producer in Africa with a share of 6 million tons per year, it still imports rice to satisfy its local needs due to production loss, especially due to rice disease. Rice blast disease is responsible for 30% loss in rice production worldwide. Therefore, it is crucial to target limiting yield damage by detecting rice crops diseases in its early stages. This paper introduces a public multispectral and RGB images dataset and a deep learning pipeline for rice plant disease detection using multi-modal data. The collected multispectral images consist of Red, Green and Near-Infrared channels and we show that using multispectral along with RGB channels as input archives a higher F1 accuracy compared to using RGB input only.
### Combining deep learning and street view imagery to map smallholder crop types
- **Authors:** Jordi Laguarta, Thomas Friedel, Sherrie Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.05930
- **Pdf link:** https://arxiv.org/pdf/2309.05930
- **Abstract**
Accurate crop type maps are an essential source of information for monitoring yield progress at scale, projecting global crop production, and planning effective policies. To date, however, crop type maps remain challenging to create in low and middle-income countries due to a lack of ground truth labels for training machine learning models. Field surveys are the gold standard in terms of accuracy but require an often-prohibitively large amount of time, money, and statistical capacity. In recent years, street-level imagery, such as Google Street View, KartaView, and Mapillary, has become available around the world. Such imagery contains rich information about crop types grown at particular locations and times. In this work, we develop an automated system to generate crop type ground references using deep learning and Google Street View imagery. The method efficiently curates a set of street view images containing crop fields, trains a model to predict crop type by utilizing weakly-labelled images from disparate out-of-domain sources, and combines predicted labels with remote sensing time series to create a wall-to-wall crop type map. We show that, in Thailand, the resulting country-wide map of rice, cassava, maize, and sugarcane achieves an accuracy of 93%. As the availability of roadside imagery expands, our pipeline provides a way to map crop types at scale around the globe, especially in underserved smallholder regions.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Real-Time Semantic Segmentation: A Brief Survey & Comparative Study in Remote Sensing
- **Authors:** Clifford Broni-Bediako, Junshi Xia, Naoto Yokoya
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.06047
- **Pdf link:** https://arxiv.org/pdf/2309.06047
- **Abstract**
Real-time semantic segmentation of remote sensing imagery is a challenging task that requires a trade-off between effectiveness and efficiency. It has many applications including tracking forest fires, detecting changes in land use and land cover, crop health monitoring, and so on. With the success of efficient deep learning methods (i.e., efficient deep neural networks) for real-time semantic segmentation in computer vision, researchers have adopted these efficient deep neural networks in remote sensing image analysis. This paper begins with a summary of the fundamental compression methods for designing efficient deep neural networks and provides a brief but comprehensive survey, outlining the recent developments in real-time semantic segmentation of remote sensing imagery. We examine several seminal efficient deep learning methods, placing them in a taxonomy based on the network architecture design approach. Furthermore, we evaluate the quality and efficiency of some existing efficient deep neural networks on a publicly available remote sensing semantic segmentation benchmark dataset, the OpenEarthMap. The experimental results of an extensive comparative study demonstrate that most of the existing efficient deep neural networks have good segmentation quality, but they suffer low inference speed (i.e., high latency rate), which may limit their capability of deployment in real-time applications of remote sensing image segmentation. We provide some insights into the current trend and future research directions for real-time semantic segmentation of remote sensing imagery.
## Keyword: RAW
### Adversarial Attacks Assessment of Salient Object Detection via Symbolic Learning
- **Authors:** Gustavo Olague, Roberto Pineda, Gerardo Ibarra-Vazquez, Matthieu Olague, Axel Martinez, Sambit Bakshi, Jonathan Vargas, Isnardo Reducindo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE)
- **Arxiv link:** https://arxiv.org/abs/2309.05900
- **Pdf link:** https://arxiv.org/pdf/2309.05900
- **Abstract**
Machine learning is at the center of mainstream technology and outperforms classical approaches to handcrafted feature design. Aside from its learning process for artificial feature extraction, it has an end-to-end paradigm from input to output, reaching outstandingly accurate results. However, security concerns about its robustness to malicious and imperceptible perturbations have drawn attention since its prediction can be changed entirely. Salient object detection is a research area where deep convolutional neural networks have proven effective but whose trustworthiness represents a significant issue requiring analysis and solutions to hackers' attacks. Brain programming is a kind of symbolic learning in the vein of good old-fashioned artificial intelligence. This work provides evidence that symbolic learning robustness is crucial in designing reliable visual attention systems since it can withstand even the most intense perturbations. We test this evolutionary computation methodology against several adversarial attacks and noise perturbations using standard databases and a real-world problem of a shorebird called the Snowy Plover portraying a visual attention task. We compare our methodology with five different deep learning approaches, proving that they do not match the symbolic paradigm regarding robustness. All neural networks suffer significant performance losses, while brain programming stands its ground and remains unaffected. Also, by studying the Snowy Plover, we remark on the importance of security in surveillance activities regarding wildlife protection and conservation.
### TSSAT: Two-Stage Statistics-Aware Transformation for Artistic Style Transfer
- **Authors:** Haibo Chen, Lei Zhao, Jun Li, Jian Yang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.06004
- **Pdf link:** https://arxiv.org/pdf/2309.06004
- **Abstract**
Artistic style transfer aims to create new artistic images by rendering a given photograph with the target artistic style. Existing methods learn styles simply based on global statistics or local patches, lacking careful consideration of the drawing process in practice. Consequently, the stylization results either fail to capture abundant and diversified local style patterns, or contain undesired semantic information of the style image and deviate from the global style distribution. To address this issue, we imitate the drawing process of humans and propose a Two-Stage Statistics-Aware Transformation (TSSAT) module, which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics (instead of local features) in a patch-wise manner, significantly improving the stylization effects. Moreover, to further enhance both content and style representations, we introduce two novel losses: an attention-based content loss and a patch-based style loss, where the former enables better content preservation by enforcing the semantic relation in the content image to be retained during stylization, and the latter focuses on increasing the local style similarity between the style and stylized images. Extensive qualitative and quantitative experiments verify the effectiveness of our method.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Wed, 13 Sep 23 - ## Keyword: events
### SoccerNet 2023 Challenges Results
- **Authors:** Anthony Cioppa, Silvio Giancola, Vladimir Somers, Floriane Magera, Xin Zhou, Hassan Mkhallati, Adrien Deliège, Jan Held, Carlos Hinojosa, Amir M. Mansourian, Pierre Miralles, Olivier Barnich, Christophe De Vleeschouwer, Alexandre Alahi, Bernard Ghanem, Marc Van Droogenbroeck, Abdullah Kamal, Adrien Maglo, Albert Clapés, Amr Abdelaziz, Artur Xarles, Astrid Orcesi, Atom Scott, Bin Liu, Byoungkwon Lim, Chen Chen, Fabian Deuser, Feng Yan, Fufu Yu, Gal Shitrit, Guanshuo Wang, Gyusik Choi, Hankyul Kim, Hao Guo, Hasby Fahrudin, Hidenari Koguchi, Håkan Ardö, Ibrahim Salah, Ido Yerushalmy, Iftikar Muhammad, Ikuma Uchida, Ishay Be'ery, Jaonary Rabarisoa, Jeongae Lee, Jiajun Fu, Jianqin Yin, Jinghang Xu, Jongho Nang, Julien Denize, Junjie Li, Junpei Zhang, Juntae Kim, Kamil Synowiec, et al. (49 additional authors not shown)
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.06006
- **Pdf link:** https://arxiv.org/pdf/2309.06006
- **Abstract**
The SoccerNet 2023 challenges were the third annual video understanding challenges organized by the SoccerNet team. For this third edition, the challenges were composed of seven vision-based tasks split into three main themes. The first theme, broadcast video understanding, is composed of three high-level tasks related to describing events occurring in the video broadcasts: (1) action spotting, focusing on retrieving all timestamps related to global actions in soccer, (2) ball action spotting, focusing on retrieving all timestamps related to the soccer ball change of state, and (3) dense video captioning, focusing on describing the broadcast with natural language and anchored timestamps. The second theme, field understanding, relates to the single task of (4) camera calibration, focusing on retrieving the intrinsic and extrinsic camera parameters from images. The third and last theme, player understanding, is composed of three low-level tasks related to extracting information about the players: (5) re-identification, focusing on retrieving the same players across multiple views, (6) multiple object tracking, focusing on tracking players and the ball through unedited video streams, and (7) jersey number recognition, focusing on recognizing the jersey number of players from tracklets. Compared to the previous editions of the SoccerNet challenges, tasks (2-3-7) are novel, including new annotations and data, task (4) was enhanced with more data and annotations, and task (6) now focuses on end-to-end approaches. More information on the tasks, challenges, and leaderboards are available on https://www.soccer-net.org. Baselines and development kits can be found on https://github.com/SoccerNet.
### JOADAA: joint online action detection and action anticipation
- **Authors:** Mohammed Guermal, Francois Bremond, Rui Dai, Abid Ali
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.06130
- **Pdf link:** https://arxiv.org/pdf/2309.06130
- **Abstract**
Action anticipation involves forecasting future actions by connecting past events to future ones. However, this reasoning ignores the real-life hierarchy of events which is considered to be composed of three main parts: past, present, and future. We argue that considering these three main parts and their dependencies could improve performance. On the other hand, online action detection is the task of predicting actions in a streaming manner. In this case, one has access only to the past and present information. Therefore, in online action detection (OAD) the existing approaches miss semantics or future information which limits their performance. To sum up, for both of these tasks, the complete set of knowledge (past-present-future) is missing, which makes it challenging to infer action dependencies, therefore having low performances. To address this limitation, we propose to fuse both tasks into a single uniform architecture. By combining action anticipation and online action detection, our approach can cover the missing dependencies of future information in online action detection. This method referred to as JOADAA, presents a uniform model that jointly performs action anticipation and online action detection. We validate our proposed model on three challenging datasets: THUMOS'14, which is a sparsely annotated dataset with one action per time step, CHARADES, and Multi-THUMOS, two densely annotated datasets with more complex scenarios. JOADAA achieves SOTA results on these benchmarks for both tasks.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Rice Plant Disease Detection and Diagnosis using Deep Convolutional Neural Networks and Multispectral Imaging
- **Authors:** Yara Ali Alnaggar, Ahmad Sebaq, Karim Amer, ElSayed Naeem, Mohamed Elhelw
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2309.05818
- **Pdf link:** https://arxiv.org/pdf/2309.05818
- **Abstract**
Rice is considered a strategic crop in Egypt as it is regularly consumed in the Egyptian people's diet. Even though Egypt is the highest rice producer in Africa with a share of 6 million tons per year, it still imports rice to satisfy its local needs due to production loss, especially due to rice disease. Rice blast disease is responsible for 30% loss in rice production worldwide. Therefore, it is crucial to target limiting yield damage by detecting rice crops diseases in its early stages. This paper introduces a public multispectral and RGB images dataset and a deep learning pipeline for rice plant disease detection using multi-modal data. The collected multispectral images consist of Red, Green and Near-Infrared channels and we show that using multispectral along with RGB channels as input archives a higher F1 accuracy compared to using RGB input only.
### Combining deep learning and street view imagery to map smallholder crop types
- **Authors:** Jordi Laguarta, Thomas Friedel, Sherrie Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.05930
- **Pdf link:** https://arxiv.org/pdf/2309.05930
- **Abstract**
Accurate crop type maps are an essential source of information for monitoring yield progress at scale, projecting global crop production, and planning effective policies. To date, however, crop type maps remain challenging to create in low and middle-income countries due to a lack of ground truth labels for training machine learning models. Field surveys are the gold standard in terms of accuracy but require an often-prohibitively large amount of time, money, and statistical capacity. In recent years, street-level imagery, such as Google Street View, KartaView, and Mapillary, has become available around the world. Such imagery contains rich information about crop types grown at particular locations and times. In this work, we develop an automated system to generate crop type ground references using deep learning and Google Street View imagery. The method efficiently curates a set of street view images containing crop fields, trains a model to predict crop type by utilizing weakly-labelled images from disparate out-of-domain sources, and combines predicted labels with remote sensing time series to create a wall-to-wall crop type map. We show that, in Thailand, the resulting country-wide map of rice, cassava, maize, and sugarcane achieves an accuracy of 93%. As the availability of roadside imagery expands, our pipeline provides a way to map crop types at scale around the globe, especially in underserved smallholder regions.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Real-Time Semantic Segmentation: A Brief Survey & Comparative Study in Remote Sensing
- **Authors:** Clifford Broni-Bediako, Junshi Xia, Naoto Yokoya
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.06047
- **Pdf link:** https://arxiv.org/pdf/2309.06047
- **Abstract**
Real-time semantic segmentation of remote sensing imagery is a challenging task that requires a trade-off between effectiveness and efficiency. It has many applications including tracking forest fires, detecting changes in land use and land cover, crop health monitoring, and so on. With the success of efficient deep learning methods (i.e., efficient deep neural networks) for real-time semantic segmentation in computer vision, researchers have adopted these efficient deep neural networks in remote sensing image analysis. This paper begins with a summary of the fundamental compression methods for designing efficient deep neural networks and provides a brief but comprehensive survey, outlining the recent developments in real-time semantic segmentation of remote sensing imagery. We examine several seminal efficient deep learning methods, placing them in a taxonomy based on the network architecture design approach. Furthermore, we evaluate the quality and efficiency of some existing efficient deep neural networks on a publicly available remote sensing semantic segmentation benchmark dataset, the OpenEarthMap. The experimental results of an extensive comparative study demonstrate that most of the existing efficient deep neural networks have good segmentation quality, but they suffer low inference speed (i.e., high latency rate), which may limit their capability of deployment in real-time applications of remote sensing image segmentation. We provide some insights into the current trend and future research directions for real-time semantic segmentation of remote sensing imagery.
## Keyword: RAW
### Adversarial Attacks Assessment of Salient Object Detection via Symbolic Learning
- **Authors:** Gustavo Olague, Roberto Pineda, Gerardo Ibarra-Vazquez, Matthieu Olague, Axel Martinez, Sambit Bakshi, Jonathan Vargas, Isnardo Reducindo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE)
- **Arxiv link:** https://arxiv.org/abs/2309.05900
- **Pdf link:** https://arxiv.org/pdf/2309.05900
- **Abstract**
Machine learning is at the center of mainstream technology and outperforms classical approaches to handcrafted feature design. Aside from its learning process for artificial feature extraction, it has an end-to-end paradigm from input to output, reaching outstandingly accurate results. However, security concerns about its robustness to malicious and imperceptible perturbations have drawn attention since its prediction can be changed entirely. Salient object detection is a research area where deep convolutional neural networks have proven effective but whose trustworthiness represents a significant issue requiring analysis and solutions to hackers' attacks. Brain programming is a kind of symbolic learning in the vein of good old-fashioned artificial intelligence. This work provides evidence that symbolic learning robustness is crucial in designing reliable visual attention systems since it can withstand even the most intense perturbations. We test this evolutionary computation methodology against several adversarial attacks and noise perturbations using standard databases and a real-world problem of a shorebird called the Snowy Plover portraying a visual attention task. We compare our methodology with five different deep learning approaches, proving that they do not match the symbolic paradigm regarding robustness. All neural networks suffer significant performance losses, while brain programming stands its ground and remains unaffected. Also, by studying the Snowy Plover, we remark on the importance of security in surveillance activities regarding wildlife protection and conservation.
### TSSAT: Two-Stage Statistics-Aware Transformation for Artistic Style Transfer
- **Authors:** Haibo Chen, Lei Zhao, Jun Li, Jian Yang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.06004
- **Pdf link:** https://arxiv.org/pdf/2309.06004
- **Abstract**
Artistic style transfer aims to create new artistic images by rendering a given photograph with the target artistic style. Existing methods learn styles simply based on global statistics or local patches, lacking careful consideration of the drawing process in practice. Consequently, the stylization results either fail to capture abundant and diversified local style patterns, or contain undesired semantic information of the style image and deviate from the global style distribution. To address this issue, we imitate the drawing process of humans and propose a Two-Stage Statistics-Aware Transformation (TSSAT) module, which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics (instead of local features) in a patch-wise manner, significantly improving the stylization effects. Moreover, to further enhance both content and style representations, we introduce two novel losses: an attention-based content loss and a patch-based style loss, where the former enables better content preservation by enforcing the semantic relation in the content image to be retained during stylization, and the latter focuses on increasing the local style similarity between the style and stylized images. Extensive qualitative and quantitative experiments verify the effectiveness of our method.
## Keyword: raw image
There is no result
|
process
|
new submissions for wed sep keyword events soccernet challenges results authors anthony cioppa silvio giancola vladimir somers floriane magera xin zhou hassan mkhallati adrien deliège jan held carlos hinojosa amir m mansourian pierre miralles olivier barnich christophe de vleeschouwer alexandre alahi bernard ghanem marc van droogenbroeck abdullah kamal adrien maglo albert clapés amr abdelaziz artur xarles astrid orcesi atom scott bin liu byoungkwon lim chen chen fabian deuser feng yan fufu yu gal shitrit guanshuo wang gyusik choi hankyul kim hao guo hasby fahrudin hidenari koguchi håkan ardö ibrahim salah ido yerushalmy iftikar muhammad ikuma uchida ishay be ery jaonary rabarisoa jeongae lee jiajun fu jianqin yin jinghang xu jongho nang julien denize junjie li junpei zhang juntae kim kamil synowiec et al additional authors not shown subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract the soccernet challenges were the third annual video understanding challenges organized by the soccernet team for this third edition the challenges were composed of seven vision based tasks split into three main themes the first theme broadcast video understanding is composed of three high level tasks related to describing events occurring in the video broadcasts action spotting focusing on retrieving all timestamps related to global actions in soccer ball action spotting focusing on retrieving all timestamps related to the soccer ball change of state and dense video captioning focusing on describing the broadcast with natural language and anchored timestamps the second theme field understanding relates to the single task of camera calibration focusing on retrieving the intrinsic and extrinsic camera parameters from images the third and last theme player understanding is composed of three low level tasks related to extracting information about the players re identification focusing on retrieving the same players across multiple views multiple object tracking focusing on tracking players and the ball through unedited video streams and jersey number recognition focusing on recognizing the jersey number of players from tracklets compared to the previous editions of the soccernet challenges tasks are novel including new annotations and data task was enhanced with more data and annotations and task now focuses on end to end approaches more information on the tasks challenges and leaderboards are available on baselines and development kits can be found on joadaa joint online action detection and action anticipation authors mohammed guermal francois bremond rui dai abid ali subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract action anticipation involves forecasting future actions by connecting past events to future ones however this reasoning ignores the real life hierarchy of events which is considered to be composed of three main parts past present and future we argue that considering these three main parts and their dependencies could improve performance on the other hand online action detection is the task of predicting actions in a streaming manner in this case one has access only to the past and present information therefore in online action detection oad the existing approaches miss semantics or future information which limits their performance to sum up for both of these tasks the complete set of knowledge past present future is missing which makes it challenging to infer action dependencies therefore having low performances to address this limitation we propose to fuse both tasks into a single uniform architecture by combining action anticipation and online action detection our approach can cover the missing dependencies of future information in online action detection this method referred to as joadaa presents a uniform model that jointly performs action anticipation and online action detection we validate our proposed model on three challenging datasets thumos which is a sparsely annotated dataset with one action per time step charades and multi thumos two densely annotated datasets with more complex scenarios joadaa achieves sota results on these benchmarks for both tasks keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp rice plant disease detection and diagnosis using deep convolutional neural networks and multispectral imaging authors yara ali alnaggar ahmad sebaq karim amer elsayed naeem mohamed elhelw subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract rice is considered a strategic crop in egypt as it is regularly consumed in the egyptian people s diet even though egypt is the highest rice producer in africa with a share of million tons per year it still imports rice to satisfy its local needs due to production loss especially due to rice disease rice blast disease is responsible for loss in rice production worldwide therefore it is crucial to target limiting yield damage by detecting rice crops diseases in its early stages this paper introduces a public multispectral and rgb images dataset and a deep learning pipeline for rice plant disease detection using multi modal data the collected multispectral images consist of red green and near infrared channels and we show that using multispectral along with rgb channels as input archives a higher accuracy compared to using rgb input only combining deep learning and street view imagery to map smallholder crop types authors jordi laguarta thomas friedel sherrie wang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract accurate crop type maps are an essential source of information for monitoring yield progress at scale projecting global crop production and planning effective policies to date however crop type maps remain challenging to create in low and middle income countries due to a lack of ground truth labels for training machine learning models field surveys are the gold standard in terms of accuracy but require an often prohibitively large amount of time money and statistical capacity in recent years street level imagery such as google street view kartaview and mapillary has become available around the world such imagery contains rich information about crop types grown at particular locations and times in this work we develop an automated system to generate crop type ground references using deep learning and google street view imagery the method efficiently curates a set of street view images containing crop fields trains a model to predict crop type by utilizing weakly labelled images from disparate out of domain sources and combines predicted labels with remote sensing time series to create a wall to wall crop type map we show that in thailand the resulting country wide map of rice cassava maize and sugarcane achieves an accuracy of as the availability of roadside imagery expands our pipeline provides a way to map crop types at scale around the globe especially in underserved smallholder regions keyword image signal processing there is no result keyword image signal process there is no result keyword compression real time semantic segmentation a brief survey comparative study in remote sensing authors clifford broni bediako junshi xia naoto yokoya subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract real time semantic segmentation of remote sensing imagery is a challenging task that requires a trade off between effectiveness and efficiency it has many applications including tracking forest fires detecting changes in land use and land cover crop health monitoring and so on with the success of efficient deep learning methods i e efficient deep neural networks for real time semantic segmentation in computer vision researchers have adopted these efficient deep neural networks in remote sensing image analysis this paper begins with a summary of the fundamental compression methods for designing efficient deep neural networks and provides a brief but comprehensive survey outlining the recent developments in real time semantic segmentation of remote sensing imagery we examine several seminal efficient deep learning methods placing them in a taxonomy based on the network architecture design approach furthermore we evaluate the quality and efficiency of some existing efficient deep neural networks on a publicly available remote sensing semantic segmentation benchmark dataset the openearthmap the experimental results of an extensive comparative study demonstrate that most of the existing efficient deep neural networks have good segmentation quality but they suffer low inference speed i e high latency rate which may limit their capability of deployment in real time applications of remote sensing image segmentation we provide some insights into the current trend and future research directions for real time semantic segmentation of remote sensing imagery keyword raw adversarial attacks assessment of salient object detection via symbolic learning authors gustavo olague roberto pineda gerardo ibarra vazquez matthieu olague axel martinez sambit bakshi jonathan vargas isnardo reducindo subjects computer vision and pattern recognition cs cv cryptography and security cs cr machine learning cs lg neural and evolutionary computing cs ne arxiv link pdf link abstract machine learning is at the center of mainstream technology and outperforms classical approaches to handcrafted feature design aside from its learning process for artificial feature extraction it has an end to end paradigm from input to output reaching outstandingly accurate results however security concerns about its robustness to malicious and imperceptible perturbations have drawn attention since its prediction can be changed entirely salient object detection is a research area where deep convolutional neural networks have proven effective but whose trustworthiness represents a significant issue requiring analysis and solutions to hackers attacks brain programming is a kind of symbolic learning in the vein of good old fashioned artificial intelligence this work provides evidence that symbolic learning robustness is crucial in designing reliable visual attention systems since it can withstand even the most intense perturbations we test this evolutionary computation methodology against several adversarial attacks and noise perturbations using standard databases and a real world problem of a shorebird called the snowy plover portraying a visual attention task we compare our methodology with five different deep learning approaches proving that they do not match the symbolic paradigm regarding robustness all neural networks suffer significant performance losses while brain programming stands its ground and remains unaffected also by studying the snowy plover we remark on the importance of security in surveillance activities regarding wildlife protection and conservation tssat two stage statistics aware transformation for artistic style transfer authors haibo chen lei zhao jun li jian yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract artistic style transfer aims to create new artistic images by rendering a given photograph with the target artistic style existing methods learn styles simply based on global statistics or local patches lacking careful consideration of the drawing process in practice consequently the stylization results either fail to capture abundant and diversified local style patterns or contain undesired semantic information of the style image and deviate from the global style distribution to address this issue we imitate the drawing process of humans and propose a two stage statistics aware transformation tssat module which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics instead of local features in a patch wise manner significantly improving the stylization effects moreover to further enhance both content and style representations we introduce two novel losses an attention based content loss and a patch based style loss where the former enables better content preservation by enforcing the semantic relation in the content image to be retained during stylization and the latter focuses on increasing the local style similarity between the style and stylized images extensive qualitative and quantitative experiments verify the effectiveness of our method keyword raw image there is no result
| 1
|
14,645
| 17,773,567,497
|
IssuesEvent
|
2021-08-30 16:16:47
|
googleapis/python-storage
|
https://api.github.com/repos/googleapis/python-storage
|
closed
|
Dependency Dashboard
|
api: storage type: process
|
This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/)
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/googleapis-common-protos-1.x -->[chore(deps): update dependency googleapis-common-protos to <1.53.1](../pull/432)
- [ ] <!-- recreate-branch=renovate/google-api-core-2.x -->[chore(deps): update dependency google-api-core to v2](../pull/558)
- [ ] <!-- recreate-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2](../pull/542)
- [ ] <!-- recreate-branch=renovate/google-resumable-media-2.x -->[chore(deps): update dependency google-resumable-media to v2](../pull/547)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/)
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/googleapis-common-protos-1.x -->[chore(deps): update dependency googleapis-common-protos to <1.53.1](../pull/432)
- [ ] <!-- recreate-branch=renovate/google-api-core-2.x -->[chore(deps): update dependency google-api-core to v2](../pull/558)
- [ ] <!-- recreate-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2](../pull/542)
- [ ] <!-- recreate-branch=renovate/google-resumable-media-2.x -->[chore(deps): update dependency google-resumable-media to v2](../pull/547)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue provides visibility into renovate updates and their statuses ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull pull pull check this box to trigger a request for renovate to run again on this repository
| 1
|
119,972
| 17,644,003,757
|
IssuesEvent
|
2021-08-20 01:26:19
|
AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
|
https://api.github.com/repos/AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
|
opened
|
CVE-2021-29529 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2021-29529 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. An attacker can trigger a heap buffer overflow in `tf.raw_ops.QuantizedResizeBilinear` by manipulating input values so that float rounding results in off-by-one error in accessing image elements. This is because the implementation(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L62-L66) computes two integers (representing the upper and lower bounds for interpolation) by ceiling and flooring a floating point value. For some values of `in`, `interpolation->upper[i]` might be smaller than `interpolation->lower[i]`. This is an issue if `interpolation->upper[i]` is capped at `in_size-1` as it means that `interpolation->lower[i]` points outside of the image. Then, in the interpolation code(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L245-L264), this would result in heap buffer overflow. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29529>CVE-2021-29529</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-29529 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-29529 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. An attacker can trigger a heap buffer overflow in `tf.raw_ops.QuantizedResizeBilinear` by manipulating input values so that float rounding results in off-by-one error in accessing image elements. This is because the implementation(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L62-L66) computes two integers (representing the upper and lower bounds for interpolation) by ceiling and flooring a floating point value. For some values of `in`, `interpolation->upper[i]` might be smaller than `interpolation->lower[i]`. This is an issue if `interpolation->upper[i]` is capped at `in_size-1` as it means that `interpolation->lower[i]` points outside of the image. Then, in the interpolation code(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L245-L264), this would result in heap buffer overflow. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29529>CVE-2021-29529</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file finalproject requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an end to end open source platform for machine learning an attacker can trigger a heap buffer overflow in tf raw ops quantizedresizebilinear by manipulating input values so that float rounding results in off by one error in accessing image elements this is because the implementation computes two integers representing the upper and lower bounds for interpolation by ceiling and flooring a floating point value for some values of in interpolation upper might be smaller than interpolation lower this is an issue if interpolation upper is capped at in size as it means that interpolation lower points outside of the image then in the interpolation code this would result in heap buffer overflow the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
| 0
|
110,025
| 23,854,240,817
|
IssuesEvent
|
2022-09-06 21:08:48
|
WordPress/openverse-catalog
|
https://api.github.com/repos/WordPress/openverse-catalog
|
opened
|
Consider setting category to 'illustration' for all svgs
|
🟩 priority: low ✨ goal: improvement 💻 aspect: code
|
## Current Situation
<!-- Describe the part of the code you think should improve -->
Related #614
Wikimedia sets the `category` for all records of filetype `svg` to "illustration". We should consider whether it's acceptable to do this for *all* providers, and if so we can move that logic to the `ImageStore` class.
## Suggested Improvement
<!-- Describe your proposed change -->
If we do this, then we can also remove Wikimedia's [extract_filetype](https://github.com/WordPress/openverse-catalog/blob/18decf9b70f409bc83809d14b7be3b65cb91d409/openverse_catalog/dags/providers/provider_api_scripts/wikimedia_commons.py#L349) method since the MediaStore will already handle extracting filetype from the media URL if `filetype` is None.
## Benefit
<!-- Describe the benefit of the change (E.g., increase test coverage, reduce running time, etc.) -->
Improved category population for more provider scripts, and allows removal of ad-hoc (and duplicated) code from Wikimedia.
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
|
1.0
|
Consider setting category to 'illustration' for all svgs - ## Current Situation
<!-- Describe the part of the code you think should improve -->
Related #614
Wikimedia sets the `category` for all records of filetype `svg` to "illustration". We should consider whether it's acceptable to do this for *all* providers, and if so we can move that logic to the `ImageStore` class.
## Suggested Improvement
<!-- Describe your proposed change -->
If we do this, then we can also remove Wikimedia's [extract_filetype](https://github.com/WordPress/openverse-catalog/blob/18decf9b70f409bc83809d14b7be3b65cb91d409/openverse_catalog/dags/providers/provider_api_scripts/wikimedia_commons.py#L349) method since the MediaStore will already handle extracting filetype from the media URL if `filetype` is None.
## Benefit
<!-- Describe the benefit of the change (E.g., increase test coverage, reduce running time, etc.) -->
Improved category population for more provider scripts, and allows removal of ad-hoc (and duplicated) code from Wikimedia.
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
|
non_process
|
consider setting category to illustration for all svgs current situation related wikimedia sets the category for all records of filetype svg to illustration we should consider whether it s acceptable to do this for all providers and if so we can move that logic to the imagestore class suggested improvement if we do this then we can also remove wikimedia s method since the mediastore will already handle extracting filetype from the media url if filetype is none benefit improved category population for more provider scripts and allows removal of ad hoc and duplicated code from wikimedia implementation 🙋 i would be interested in implementing this feature
| 0
|
253,725
| 27,300,814,243
|
IssuesEvent
|
2023-02-24 01:40:18
|
panasalap/linux-4.19.72_1
|
https://api.github.com/repos/panasalap/linux-4.19.72_1
|
closed
|
WS-2021-0462 (Medium) detected in linux-yoctov5.4.51 - autoclosed
|
security vulnerability
|
## WS-2021-0462 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Linux/Kernel is vulnerable to check ctr->cnr to avoid array index out of bound in drivers/isdn/capi/kcapi.c
<p>Publish Date: 2021-11-29
<p>URL: <a href=https://github.com/gregkh/linux/commit/1f3e2e97c003f80c4b087092b225c8787ff91e4d>WS-2021-0462</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1002166">https://osv.dev/vulnerability/UVI-2021-1002166</a></p>
<p>Release Date: 2021-11-29</p>
<p>Fix Resolution: Linux/Kernel - v4.4.290, v4.9.288, v4.14.253, v4.19.214, v5.4.156, v5.10.76, v5.14.15, v5.15-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0462 (Medium) detected in linux-yoctov5.4.51 - autoclosed - ## WS-2021-0462 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Linux/Kernel is vulnerable to check ctr->cnr to avoid array index out of bound in drivers/isdn/capi/kcapi.c
<p>Publish Date: 2021-11-29
<p>URL: <a href=https://github.com/gregkh/linux/commit/1f3e2e97c003f80c4b087092b225c8787ff91e4d>WS-2021-0462</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1002166">https://osv.dev/vulnerability/UVI-2021-1002166</a></p>
<p>Release Date: 2021-11-29</p>
<p>Fix Resolution: Linux/Kernel - v4.4.290, v4.9.288, v4.14.253, v4.19.214, v5.4.156, v5.10.76, v5.14.15, v5.15-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in linux autoclosed ws medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers isdn capi kcapi c drivers isdn capi kcapi c vulnerability details in linux kernel is vulnerable to check ctr cnr to avoid array index out of bound in drivers isdn capi kcapi c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux kernel step up your open source security game with mend
| 0
|
121,237
| 25,941,646,322
|
IssuesEvent
|
2022-12-16 19:02:51
|
xygeni/xygeni-goat
|
https://api.github.com/repos/xygeni/xygeni-goat
|
opened
|
Add initial curated list of bad code
|
bad code
|
The goal is to add sources with intentionally flawed code, configurations, dependencies graph, IaC templates, etc. for learning about Software Supply Chain Security, and for running security tools.
### Location of files
The initial set of bad code will be added under source/KIND, with KIND in { `secrets`, `misconfigs`, `bad_deps`, `iac_flaws` or `code_tamper` }. Under each directory, arbitrary directories could be defined for holding the bad code (for example each containing the code for a particular kind of defect).
Please note that directories may need to be added so the files are properly scanned, so for example GitHub workflows are placed under `.github/workflows`, dependency descripts like package.json or package-lock.json need to have the exact name for the scanner to recognize them, etc.
Examples:
- `source/secrets/aws/aws_key.bad.json` (for this case, probably there is no need to add an `aws_key.bad.json`)
- `source/bad_deps/scripts/bad/yarn.lock` (and corresponding package.json) and `source/bad_deps/scripts/good/yarn.lock` (and corresponding package.json) - note that bad and good directories instead of filename suffixes are used here, to preserve the file names.
- `source/minconfigs/scriptinjection/.github/workflows/inline-script-injection.bad.yml` and corresponding `inline-script-injection.good.yml`. Note that .github/workflows should be used so the GitHub workflow YAML files are recognized.
- `source/iac/terraform/aws/alb_protected_by_waf/bad/main.tf` (and others in same directory as needed for a coherent Terraform configuration) vs. `source/iac/terraform/aws/alb_protected_by_waf/good/main.tf`. All files belonging to the same configuration must be placed under the same directory.
### Expections file - expected.yaml
An `expected.yaml` file could be added for each issue example, with two lists: `fail` with the items that have security issues, `pass` for the examples with fixed code. Each item listed could be the resource ID in IaC files, a path in the hierarchy of a configuration file like a JSON or YAML, line number or lines range. Anything helping the user to locate where is the problem, even to use the `expected.yaml` to automate verification of results from a scanner.
|
1.0
|
Add initial curated list of bad code - The goal is to add sources with intentionally flawed code, configurations, dependencies graph, IaC templates, etc. for learning about Software Supply Chain Security, and for running security tools.
### Location of files
The initial set of bad code will be added under source/KIND, with KIND in { `secrets`, `misconfigs`, `bad_deps`, `iac_flaws` or `code_tamper` }. Under each directory, arbitrary directories could be defined for holding the bad code (for example each containing the code for a particular kind of defect).
Please note that directories may need to be added so the files are properly scanned, so for example GitHub workflows are placed under `.github/workflows`, dependency descripts like package.json or package-lock.json need to have the exact name for the scanner to recognize them, etc.
Examples:
- `source/secrets/aws/aws_key.bad.json` (for this case, probably there is no need to add an `aws_key.bad.json`)
- `source/bad_deps/scripts/bad/yarn.lock` (and corresponding package.json) and `source/bad_deps/scripts/good/yarn.lock` (and corresponding package.json) - note that bad and good directories instead of filename suffixes are used here, to preserve the file names.
- `source/minconfigs/scriptinjection/.github/workflows/inline-script-injection.bad.yml` and corresponding `inline-script-injection.good.yml`. Note that .github/workflows should be used so the GitHub workflow YAML files are recognized.
- `source/iac/terraform/aws/alb_protected_by_waf/bad/main.tf` (and others in same directory as needed for a coherent Terraform configuration) vs. `source/iac/terraform/aws/alb_protected_by_waf/good/main.tf`. All files belonging to the same configuration must be placed under the same directory.
### Expections file - expected.yaml
An `expected.yaml` file could be added for each issue example, with two lists: `fail` with the items that have security issues, `pass` for the examples with fixed code. Each item listed could be the resource ID in IaC files, a path in the hierarchy of a configuration file like a JSON or YAML, line number or lines range. Anything helping the user to locate where is the problem, even to use the `expected.yaml` to automate verification of results from a scanner.
|
non_process
|
add initial curated list of bad code the goal is to add sources with intentionally flawed code configurations dependencies graph iac templates etc for learning about software supply chain security and for running security tools location of files the initial set of bad code will be added under source kind with kind in secrets misconfigs bad deps iac flaws or code tamper under each directory arbitrary directories could be defined for holding the bad code for example each containing the code for a particular kind of defect please note that directories may need to be added so the files are properly scanned so for example github workflows are placed under github workflows dependency descripts like package json or package lock json need to have the exact name for the scanner to recognize them etc examples source secrets aws aws key bad json for this case probably there is no need to add an aws key bad json source bad deps scripts bad yarn lock and corresponding package json and source bad deps scripts good yarn lock and corresponding package json note that bad and good directories instead of filename suffixes are used here to preserve the file names source minconfigs scriptinjection github workflows inline script injection bad yml and corresponding inline script injection good yml note that github workflows should be used so the github workflow yaml files are recognized source iac terraform aws alb protected by waf bad main tf and others in same directory as needed for a coherent terraform configuration vs source iac terraform aws alb protected by waf good main tf all files belonging to the same configuration must be placed under the same directory expections file expected yaml an expected yaml file could be added for each issue example with two lists fail with the items that have security issues pass for the examples with fixed code each item listed could be the resource id in iac files a path in the hierarchy of a configuration file like a json or yaml line number or lines range anything helping the user to locate where is the problem even to use the expected yaml to automate verification of results from a scanner
| 0
|
17,973
| 23,984,561,074
|
IssuesEvent
|
2022-09-13 17:51:54
|
googleapis/google-cloud-node
|
https://api.github.com/repos/googleapis/google-cloud-node
|
opened
|
Configure blunderbuss and CODEOWNERS on this repository
|
type: process
|
This repo will contain hundreds of client libraries and will want some help assigning issues and PRs to the correct reviewers.
|
1.0
|
Configure blunderbuss and CODEOWNERS on this repository - This repo will contain hundreds of client libraries and will want some help assigning issues and PRs to the correct reviewers.
|
process
|
configure blunderbuss and codeowners on this repository this repo will contain hundreds of client libraries and will want some help assigning issues and prs to the correct reviewers
| 1
|
18,770
| 24,674,394,274
|
IssuesEvent
|
2022-10-18 15:51:04
|
deepset-ai/haystack
|
https://api.github.com/repos/deepset-ai/haystack
|
closed
|
Page number in document meta data not correct
|
type:bug topic:file_converter topic:preprocessing
|
**Describe the bug**
The new "add page number" features, implemented in #2932, seem to have a bug in combination with the `PDFToTextConverter`
**Error message**
Wrong page number in metadata
**To Reproduce**
Use the following haystack components:
```python
elastic_search_document_store = ElasticsearchDocumentStore(...)
retriever = BM25Retriever(document_store=elastic_search_document_store, all_terms_must_match=False)
converter = PDFToTextConverter(remove_numeric_tables=True, valid_languages=["de", "en"])
pre_processor = PreProcessor(split_length=100, add_page_number=True) # here we use the new page number flag
reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2", use_gpu=False)
```
Now we upload, convert and pre-process the document [tesla_annual_report.pdf](https://github.com/deepset-ai/haystack/files/9430730/tesla_annual_report.pdf)
```python
docs: list[Document] = converter.convert(Path(tmp_file.name), meta={'filename': filename})
docs_pre_processed: list[Document] = pre_processor.process(docs)
elastic_search_document_store.write_documents(docs_pre_processed)
```
Now we search for the search string *'financial statement'*:
```python
search_query = "financial statement"
result_documents: list[Document] = retriever.retrieve(query=search_query, top_k=30)
answers: list[Answer] = self.reader.predict(search_query, result_documents)['answers']
```
if we have a look at the top answer (and the corresponding `Document`) we see the answer
```
'reliability
of financial reporting and the preparation of financial statements for external purposes in accordance with generally accepted
accounting principles'
```
and the meta data **page** in the `Document` is **45**. If we though look into the pdf-document and search manually for the answer, we find the answer on page **48**.
The bug appears frequently, with a number of other pdf.
**FAQ Check**
- [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
**System:**
- OS: *Ubuntu 20.04*
- GPU/CPU: *both*
- Haystack version (commit or version number): *master, c0fbe45c029247428cee3823292bd574b1846c38*
- DocumentStore: *see "To Reproduce"*
- Reader: *see "To Reproduce"*
- Retriever: *see "To Reproduce"*
|
1.0
|
Page number in document meta data not correct - **Describe the bug**
The new "add page number" features, implemented in #2932, seem to have a bug in combination with the `PDFToTextConverter`
**Error message**
Wrong page number in metadata
**To Reproduce**
Use the following haystack components:
```python
elastic_search_document_store = ElasticsearchDocumentStore(...)
retriever = BM25Retriever(document_store=elastic_search_document_store, all_terms_must_match=False)
converter = PDFToTextConverter(remove_numeric_tables=True, valid_languages=["de", "en"])
pre_processor = PreProcessor(split_length=100, add_page_number=True) # here we use the new page number flag
reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2", use_gpu=False)
```
Now we upload, convert and pre-process the document [tesla_annual_report.pdf](https://github.com/deepset-ai/haystack/files/9430730/tesla_annual_report.pdf)
```python
docs: list[Document] = converter.convert(Path(tmp_file.name), meta={'filename': filename})
docs_pre_processed: list[Document] = pre_processor.process(docs)
elastic_search_document_store.write_documents(docs_pre_processed)
```
Now we search for the search string *'financial statement'*:
```python
search_query = "financial statement"
result_documents: list[Document] = retriever.retrieve(query=search_query, top_k=30)
answers: list[Answer] = self.reader.predict(search_query, result_documents)['answers']
```
if we have a look at the top answer (and the corresponding `Document`) we see the answer
```
'reliability
of financial reporting and the preparation of financial statements for external purposes in accordance with generally accepted
accounting principles'
```
and the meta data **page** in the `Document` is **45**. If we though look into the pdf-document and search manually for the answer, we find the answer on page **48**.
The bug appears frequently, with a number of other pdf.
**FAQ Check**
- [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
**System:**
- OS: *Ubuntu 20.04*
- GPU/CPU: *both*
- Haystack version (commit or version number): *master, c0fbe45c029247428cee3823292bd574b1846c38*
- DocumentStore: *see "To Reproduce"*
- Reader: *see "To Reproduce"*
- Retriever: *see "To Reproduce"*
|
process
|
page number in document meta data not correct describe the bug the new add page number features implemented in seem to have a bug in combination with the pdftotextconverter error message wrong page number in metadata to reproduce use the following haystack components python elastic search document store elasticsearchdocumentstore retriever document store elastic search document store all terms must match false converter pdftotextconverter remove numeric tables true valid languages pre processor preprocessor split length add page number true here we use the new page number flag reader farmreader model name or path deepset roberta base use gpu false now we upload convert and pre process the document python docs list converter convert path tmp file name meta filename filename docs pre processed list pre processor process docs elastic search document store write documents docs pre processed now we search for the search string financial statement python search query financial statement result documents list retriever retrieve query search query top k answers list self reader predict search query result documents if we have a look at the top answer and the corresponding document we see the answer reliability of financial reporting and the preparation of financial statements for external purposes in accordance with generally accepted accounting principles and the meta data page in the document is if we though look into the pdf document and search manually for the answer we find the answer on page the bug appears frequently with a number of other pdf faq check have you had a look at system os ubuntu gpu cpu both haystack version commit or version number master documentstore see to reproduce reader see to reproduce retriever see to reproduce
| 1
|
19,079
| 25,119,464,349
|
IssuesEvent
|
2022-11-09 06:39:39
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
opened
|
[Connector] Do not fail job when PulsarAdmin API fails when getting the data.
|
compute/data-processing
|
Currently, Pulsar connector periodically queries the admin api to detect partition changes. However, this can cause the pipeline to fail if the admin api fail.
@weixiangchen Reported a admin 502 when running pipeline for longer peirods (30 minutes).
We should discuss whether we want to
1) fail the job
2) restart the job
3) report the exception and go on
when similiar 502 exception happens.
|
1.0
|
[Connector] Do not fail job when PulsarAdmin API fails when getting the data. - Currently, Pulsar connector periodically queries the admin api to detect partition changes. However, this can cause the pipeline to fail if the admin api fail.
@weixiangchen Reported a admin 502 when running pipeline for longer peirods (30 minutes).
We should discuss whether we want to
1) fail the job
2) restart the job
3) report the exception and go on
when similiar 502 exception happens.
|
process
|
do not fail job when pulsaradmin api fails when getting the data currently pulsar connector periodically queries the admin api to detect partition changes however this can cause the pipeline to fail if the admin api fail weixiangchen reported a admin when running pipeline for longer peirods minutes we should discuss whether we want to fail the job restart the job report the exception and go on when similiar exception happens
| 1
|
12,914
| 15,287,552,000
|
IssuesEvent
|
2021-02-23 15:53:25
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
vdc user to be able to set autoscale limit
|
process_wontfix type_feature
|
the user may need to specify an autoscale limit so they don't grow indefinitely
|
1.0
|
vdc user to be able to set autoscale limit - the user may need to specify an autoscale limit so they don't grow indefinitely
|
process
|
vdc user to be able to set autoscale limit the user may need to specify an autoscale limit so they don t grow indefinitely
| 1
|
107,836
| 9,231,248,308
|
IssuesEvent
|
2019-03-13 01:29:26
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
ICE Type parameter out of range when substituting
|
A-typesystem E-needstest I-ICE T-compiler
|
Both nightly and stable ICE on this code:
```rust
use once_cell::sync::OnceCell;
use std::collections::HashMap;
type Cache<K, T> = OnceCell<HashMap<K, T>>;
trait Provider<K, T> {
fn new_cache() -> Cache<K, T> {
OnceCell::INIT
}
}
struct Fib;
impl Provider<u32, u32> for Fib {}
fn main() {
let _ = Fib::new_cache();
}
```
I expected to see this happen:
This should either compile or throw an Error explaining why it is invalid.
Instead, this happened:
Both `cargo build` and `cargo +nightly build` throw an ICE.
## Meta
installed dependencies:
```toml
once_cell = "0.1.7"
```
`rustc --version --verbose`:
rustc 1.32.0 (9fda7c223 2019-01-16)
binary: rustc
commit-hash: 9fda7c2237db910e41d6a712e9a2139b352e558b
commit-date: 2019-01-16
host: x86_64-unknown-linux-gnu
release: 1.32.0
LLVM version: 8.0
`rustc +nightly --version --verbose`:
rustc 1.33.0-nightly (0c0c58528 2019-01-19)
binary: rustc
commit-hash: 0c0c585281740aba4c91cbfd385f5e1fbe313d11
commit-date: 2019-01-19
host: x86_64-unknown-linux-gnu
release: 1.33.0-nightly
LLVM version: 8.0
Backtrace:
```
error: internal compiler error: src/librustc/ty/subst.rs:491: Type parameter `K/#1` (K/1) out of range when substituting (root type=Some(std::option::Option<std::collections::HashMap<K, T>>)) substs=[std::collections::HashMap<K, T>]
thread 'main' panicked at 'Box<Any>', src/librustc_errors/lib.rs:538:9
stack backtrace:
0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace
at src/libstd/sys/unix/backtrace/tracing/gcc_s.rs:49
1: std::sys_common::backtrace::_print
at src/libstd/sys_common/backtrace.rs:71
2: std::panicking::default_hook::{{closure}}
at src/libstd/sys_common/backtrace.rs:59
at src/libstd/panicking.rs:211
3: std::panicking::default_hook
at src/libstd/panicking.rs:227
4: rustc::util::common::panic_hook
5: std::panicking::rust_panic_with_hook
at src/libstd/panicking.rs:495
6: std::panicking::begin_panic
7: rustc_errors::Handler::span_bug
8: rustc::util::bug::opt_span_bug_fmt::{{closure}}
9: rustc::ty::context::tls::with_opt::{{closure}}
10: rustc::ty::context::tls::with_context_opt
11: rustc::ty::context::tls::with_opt
12: rustc::util::bug::opt_span_bug_fmt
13: rustc::util::bug::span_bug_fmt
14: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty
15: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter
16: rustc::ty::fold::TypeFoldable::fold_with
17: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty
18: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter
19: rustc::ty::fold::TypeFoldable::fold_with
20: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty
21: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter
22: rustc::ty::fold::TypeFoldable::fold_with
23: rustc::traits::codegen::<impl rustc::ty::context::TyCtxt<'a, 'tcx, 'tcx>>::subst_and_normalize_erasing_regions
24: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::resolve
25: rustc_mir::interpret::step::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::run
26: rustc_mir::const_eval::eval_body_using_ecx
27: rustc_mir::const_eval::const_eval_raw_provider
28: rustc::ty::query::__query_compute::const_eval_raw
29: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::const_eval_raw<'tcx>>::compute
30: rustc::ty::context::tls::with_context
31: rustc::dep_graph::graph::DepGraph::with_task_impl
32: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start
33: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job
34: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query
35: rustc::ty::query::TyCtxtAt::const_eval_raw
36: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_eval_raw
37: rustc_mir::interpret::operand::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_value_to_op
38: rustc_mir::const_eval::const_to_op
39: rustc_mir::transform::const_prop::ConstPropagator::eval_constant
40: <rustc_mir::transform::const_prop::ConstPropagator<'b, 'a, 'tcx> as rustc::mir::visit::Visitor<'tcx>>::visit_statement
41: <rustc_mir::transform::const_prop::ConstProp as rustc_mir::transform::MirPass>::run_pass
42: rustc_mir::transform::run_passes::{{closure}}
43: rustc_mir::transform::run_passes
44: rustc_mir::transform::optimized_mir
45: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::optimized_mir<'tcx>>::compute
46: rustc::ty::context::tls::with_context
47: rustc::dep_graph::graph::DepGraph::with_task_impl
48: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start
49: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job
50: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::try_get_query
51: rustc::ty::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::instance_mir
52: rustc_mir::monomorphize::collector::collect_items_rec
53: rustc_mir::monomorphize::collector::collect_items_rec
54: rustc_mir::monomorphize::collector::collect_crate_mono_items::{{closure}}
55: rustc::util::common::time
56: rustc_mir::monomorphize::collector::collect_crate_mono_items
57: rustc::util::common::time
58: rustc_mir::monomorphize::partitioning::collect_and_partition_mono_items
59: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::collect_and_partition_mono_items<'tcx>>::compute
60: rustc::ty::context::tls::with_context
61: rustc::dep_graph::graph::DepGraph::with_task_impl
62: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start
63: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job
64: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query
65: rustc_codegen_ssa::base::codegen_crate
66: <rustc_codegen_llvm::LlvmCodegenBackend as rustc_codegen_utils::codegen_backend::CodegenBackend>::codegen_crate
67: rustc::util::common::time
68: rustc_driver::driver::phase_4_codegen
69: rustc_driver::driver::compile_input::{{closure}}
70: rustc::ty::context::tls::enter_context
71: <std::thread::local::LocalKey<T>>::with
72: rustc::ty::context::TyCtxt::create_and_enter
73: rustc_driver::driver::compile_input
74: rustc_driver::run_compiler_with_pool
75: <scoped_tls::ScopedKey<T>>::set
76: rustc_driver::run_compiler
77: rustc_driver::monitor::{{closure}}
78: __rust_maybe_catch_panic
at src/libpanic_unwind/lib.rs:102
79: rustc_driver::run
80: rustc_driver::main
81: std::rt::lang_start::{{closure}}
82: std::panicking::try::do_call
at src/libstd/rt.rs:59
at src/libstd/panicking.rs:310
83: __rust_maybe_catch_panic
at src/libpanic_unwind/lib.rs:102
84: std::rt::lang_start_internal
at src/libstd/panicking.rs:289
at src/libstd/panic.rs:398
at src/libstd/rt.rs:58
85: main
86: __libc_start_main
87: <unknown>
query stack during panic:
#0 [const_eval_raw] const-evaluating `<once_cell::imp::OnceCell<T>>::INIT`
--> src/main.rs:8:9
|
8 | OnceCell::INIT
| ^^^^^^^^^^^^^^
#1 [optimized_mir] processing `Provider::new_cache`
#2 [collect_and_partition_mono_items] collect_and_partition_mono_items
end of query stack
```
Note:
There are multiple issues with similar titles, sorry if this is a duplicate.
Most other issues I could find seem to only affect nightly, this crashes on stable, too. #57700 panics during `mir_borrowck`, this in `optimized_mir`.
|
1.0
|
ICE Type parameter out of range when substituting - Both nightly and stable ICE on this code:
```rust
use once_cell::sync::OnceCell;
use std::collections::HashMap;
type Cache<K, T> = OnceCell<HashMap<K, T>>;
trait Provider<K, T> {
fn new_cache() -> Cache<K, T> {
OnceCell::INIT
}
}
struct Fib;
impl Provider<u32, u32> for Fib {}
fn main() {
let _ = Fib::new_cache();
}
```
I expected to see this happen:
This should either compile or throw an Error explaining why it is invalid.
Instead, this happened:
Both `cargo build` and `cargo +nightly build` throw an ICE.
## Meta
installed dependencies:
```toml
once_cell = "0.1.7"
```
`rustc --version --verbose`:
rustc 1.32.0 (9fda7c223 2019-01-16)
binary: rustc
commit-hash: 9fda7c2237db910e41d6a712e9a2139b352e558b
commit-date: 2019-01-16
host: x86_64-unknown-linux-gnu
release: 1.32.0
LLVM version: 8.0
`rustc +nightly --version --verbose`:
rustc 1.33.0-nightly (0c0c58528 2019-01-19)
binary: rustc
commit-hash: 0c0c585281740aba4c91cbfd385f5e1fbe313d11
commit-date: 2019-01-19
host: x86_64-unknown-linux-gnu
release: 1.33.0-nightly
LLVM version: 8.0
Backtrace:
```
error: internal compiler error: src/librustc/ty/subst.rs:491: Type parameter `K/#1` (K/1) out of range when substituting (root type=Some(std::option::Option<std::collections::HashMap<K, T>>)) substs=[std::collections::HashMap<K, T>]
thread 'main' panicked at 'Box<Any>', src/librustc_errors/lib.rs:538:9
stack backtrace:
0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace
at src/libstd/sys/unix/backtrace/tracing/gcc_s.rs:49
1: std::sys_common::backtrace::_print
at src/libstd/sys_common/backtrace.rs:71
2: std::panicking::default_hook::{{closure}}
at src/libstd/sys_common/backtrace.rs:59
at src/libstd/panicking.rs:211
3: std::panicking::default_hook
at src/libstd/panicking.rs:227
4: rustc::util::common::panic_hook
5: std::panicking::rust_panic_with_hook
at src/libstd/panicking.rs:495
6: std::panicking::begin_panic
7: rustc_errors::Handler::span_bug
8: rustc::util::bug::opt_span_bug_fmt::{{closure}}
9: rustc::ty::context::tls::with_opt::{{closure}}
10: rustc::ty::context::tls::with_context_opt
11: rustc::ty::context::tls::with_opt
12: rustc::util::bug::opt_span_bug_fmt
13: rustc::util::bug::span_bug_fmt
14: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty
15: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter
16: rustc::ty::fold::TypeFoldable::fold_with
17: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty
18: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter
19: rustc::ty::fold::TypeFoldable::fold_with
20: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty
21: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter
22: rustc::ty::fold::TypeFoldable::fold_with
23: rustc::traits::codegen::<impl rustc::ty::context::TyCtxt<'a, 'tcx, 'tcx>>::subst_and_normalize_erasing_regions
24: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::resolve
25: rustc_mir::interpret::step::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::run
26: rustc_mir::const_eval::eval_body_using_ecx
27: rustc_mir::const_eval::const_eval_raw_provider
28: rustc::ty::query::__query_compute::const_eval_raw
29: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::const_eval_raw<'tcx>>::compute
30: rustc::ty::context::tls::with_context
31: rustc::dep_graph::graph::DepGraph::with_task_impl
32: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start
33: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job
34: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query
35: rustc::ty::query::TyCtxtAt::const_eval_raw
36: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_eval_raw
37: rustc_mir::interpret::operand::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_value_to_op
38: rustc_mir::const_eval::const_to_op
39: rustc_mir::transform::const_prop::ConstPropagator::eval_constant
40: <rustc_mir::transform::const_prop::ConstPropagator<'b, 'a, 'tcx> as rustc::mir::visit::Visitor<'tcx>>::visit_statement
41: <rustc_mir::transform::const_prop::ConstProp as rustc_mir::transform::MirPass>::run_pass
42: rustc_mir::transform::run_passes::{{closure}}
43: rustc_mir::transform::run_passes
44: rustc_mir::transform::optimized_mir
45: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::optimized_mir<'tcx>>::compute
46: rustc::ty::context::tls::with_context
47: rustc::dep_graph::graph::DepGraph::with_task_impl
48: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start
49: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job
50: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::try_get_query
51: rustc::ty::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::instance_mir
52: rustc_mir::monomorphize::collector::collect_items_rec
53: rustc_mir::monomorphize::collector::collect_items_rec
54: rustc_mir::monomorphize::collector::collect_crate_mono_items::{{closure}}
55: rustc::util::common::time
56: rustc_mir::monomorphize::collector::collect_crate_mono_items
57: rustc::util::common::time
58: rustc_mir::monomorphize::partitioning::collect_and_partition_mono_items
59: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::collect_and_partition_mono_items<'tcx>>::compute
60: rustc::ty::context::tls::with_context
61: rustc::dep_graph::graph::DepGraph::with_task_impl
62: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start
63: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job
64: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query
65: rustc_codegen_ssa::base::codegen_crate
66: <rustc_codegen_llvm::LlvmCodegenBackend as rustc_codegen_utils::codegen_backend::CodegenBackend>::codegen_crate
67: rustc::util::common::time
68: rustc_driver::driver::phase_4_codegen
69: rustc_driver::driver::compile_input::{{closure}}
70: rustc::ty::context::tls::enter_context
71: <std::thread::local::LocalKey<T>>::with
72: rustc::ty::context::TyCtxt::create_and_enter
73: rustc_driver::driver::compile_input
74: rustc_driver::run_compiler_with_pool
75: <scoped_tls::ScopedKey<T>>::set
76: rustc_driver::run_compiler
77: rustc_driver::monitor::{{closure}}
78: __rust_maybe_catch_panic
at src/libpanic_unwind/lib.rs:102
79: rustc_driver::run
80: rustc_driver::main
81: std::rt::lang_start::{{closure}}
82: std::panicking::try::do_call
at src/libstd/rt.rs:59
at src/libstd/panicking.rs:310
83: __rust_maybe_catch_panic
at src/libpanic_unwind/lib.rs:102
84: std::rt::lang_start_internal
at src/libstd/panicking.rs:289
at src/libstd/panic.rs:398
at src/libstd/rt.rs:58
85: main
86: __libc_start_main
87: <unknown>
query stack during panic:
#0 [const_eval_raw] const-evaluating `<once_cell::imp::OnceCell<T>>::INIT`
--> src/main.rs:8:9
|
8 | OnceCell::INIT
| ^^^^^^^^^^^^^^
#1 [optimized_mir] processing `Provider::new_cache`
#2 [collect_and_partition_mono_items] collect_and_partition_mono_items
end of query stack
```
Note:
There are multiple issues with similar titles, sorry if this is a duplicate.
Most other issues I could find seem to only affect nightly, this crashes on stable, too. #57700 panics during `mir_borrowck`, this in `optimized_mir`.
|
non_process
|
ice type parameter out of range when substituting both nightly and stable ice on this code rust use once cell sync oncecell use std collections hashmap type cache oncecell trait provider fn new cache cache oncecell init struct fib impl provider for fib fn main let fib new cache i expected to see this happen this should either compile or throw an error explaining why it is invalid instead this happened both cargo build and cargo nightly build throw an ice meta installed dependencies toml once cell rustc version verbose rustc binary rustc commit hash commit date host unknown linux gnu release llvm version rustc nightly version verbose rustc nightly binary rustc commit hash commit date host unknown linux gnu release nightly llvm version backtrace error internal compiler error src librustc ty subst rs type parameter k k out of range when substituting root type some std option option substs thread main panicked at box src librustc errors lib rs stack backtrace std sys unix backtrace tracing imp unwind backtrace at src libstd sys unix backtrace tracing gcc s rs std sys common backtrace print at src libstd sys common backtrace rs std panicking default hook closure at src libstd sys common backtrace rs at src libstd panicking rs std panicking default hook at src libstd panicking rs rustc util common panic hook std panicking rust panic with hook at src libstd panicking rs std panicking begin panic rustc errors handler span bug rustc util bug opt span bug fmt closure rustc ty context tls with opt closure rustc ty context tls with context opt rustc ty context tls with opt rustc util bug opt span bug fmt rustc util bug span bug fmt as rustc ty fold typefolder fold ty as core iter traits fromiterator item from iter rustc ty fold typefoldable fold with as rustc ty fold typefolder fold ty as core iter traits fromiterator item from iter rustc ty fold typefoldable fold with as rustc ty fold typefolder fold ty as core iter traits fromiterator item from iter rustc ty fold typefoldable fold with rustc traits codegen subst and normalize erasing regions resolve rustc mir interpret step run rustc mir const eval eval body using ecx rustc mir const eval const eval raw provider rustc ty query query compute const eval raw rustc ty query for rustc ty query queries const eval raw compute rustc ty context tls with context rustc dep graph graph depgraph with task impl start rustc ty query plumbing force query with job rustc ty query plumbing get query rustc ty query tyctxtat const eval raw const eval raw rustc mir interpret operand const value to op rustc mir const eval const to op rustc mir transform const prop constpropagator eval constant as rustc mir visit visitor visit statement run pass rustc mir transform run passes closure rustc mir transform run passes rustc mir transform optimized mir rustc ty query for rustc ty query queries optimized mir compute rustc ty context tls with context rustc dep graph graph depgraph with task impl start rustc ty query plumbing force query with job rustc ty query plumbing try get query rustc ty instance mir rustc mir monomorphize collector collect items rec rustc mir monomorphize collector collect items rec rustc mir monomorphize collector collect crate mono items closure rustc util common time rustc mir monomorphize collector collect crate mono items rustc util common time rustc mir monomorphize partitioning collect and partition mono items rustc ty query for rustc ty query queries collect and partition mono items compute rustc ty context tls with context rustc dep graph graph depgraph with task impl start rustc ty query plumbing force query with job rustc ty query plumbing get query rustc codegen ssa base codegen crate codegen crate rustc util common time rustc driver driver phase codegen rustc driver driver compile input closure rustc ty context tls enter context with rustc ty context tyctxt create and enter rustc driver driver compile input rustc driver run compiler with pool set rustc driver run compiler rustc driver monitor closure rust maybe catch panic at src libpanic unwind lib rs rustc driver run rustc driver main std rt lang start closure std panicking try do call at src libstd rt rs at src libstd panicking rs rust maybe catch panic at src libpanic unwind lib rs std rt lang start internal at src libstd panicking rs at src libstd panic rs at src libstd rt rs main libc start main query stack during panic const evaluating init src main rs oncecell init processing provider new cache collect and partition mono items end of query stack note there are multiple issues with similar titles sorry if this is a duplicate most other issues i could find seem to only affect nightly this crashes on stable too panics during mir borrowck this in optimized mir
| 0
|
621,590
| 19,592,255,948
|
IssuesEvent
|
2022-01-05 14:13:53
|
WordPress/openverse-frontend
|
https://api.github.com/repos/WordPress/openverse-frontend
|
closed
|
Implement redesign and new components
|
🟧 priority: high ✨ goal: improvement 🕹 aspect: interface
|
This is a meta ticket to track the ongoing redesign efforts in Openverse. This ticket is very much a work in progress and is currently being used to track missing, ongoing issues with the redesign.
## Design Todos
### General
- [ ] #256
- [x] #273
- [ ] #349
- [ ] Simplified homepage (no featured collections yet)
- [ ] Redesigned content pages (anything text-heavy like `/sources` and so on)
- These shouldn't need much, just a refresh with the latest colors, sizing, and so on
### Mature content / copyright claims
- [ ] Add "report this {image|audio}" flow to the single result pages
<img width="372" alt="CleanShot 2021-10-29 at 10 34 29@2x" src="https://user-images.githubusercontent.com/6351754/139453661-f4270159-9a87-4b87-a82f-f8371e4e0d69.png">
- [ ] Add back filter to enable mature content (this can go in the filter sidebar)
- [ ] The reporting modal (needs to be updated in the new popover style
<img width="385" alt="CleanShot 2021-10-29 at 10 31 05@2x" src="https://user-images.githubusercontent.com/6351754/139453202-b603c027-070f-4560-a843-52b3d0541304.png">)
|
1.0
|
Implement redesign and new components - This is a meta ticket to track the ongoing redesign efforts in Openverse. This ticket is very much a work in progress and is currently being used to track missing, ongoing issues with the redesign.
## Design Todos
### General
- [ ] #256
- [x] #273
- [ ] #349
- [ ] Simplified homepage (no featured collections yet)
- [ ] Redesigned content pages (anything text-heavy like `/sources` and so on)
- These shouldn't need much, just a refresh with the latest colors, sizing, and so on
### Mature content / copyright claims
- [ ] Add "report this {image|audio}" flow to the single result pages
<img width="372" alt="CleanShot 2021-10-29 at 10 34 29@2x" src="https://user-images.githubusercontent.com/6351754/139453661-f4270159-9a87-4b87-a82f-f8371e4e0d69.png">
- [ ] Add back filter to enable mature content (this can go in the filter sidebar)
- [ ] The reporting modal (needs to be updated in the new popover style
<img width="385" alt="CleanShot 2021-10-29 at 10 31 05@2x" src="https://user-images.githubusercontent.com/6351754/139453202-b603c027-070f-4560-a843-52b3d0541304.png">)
|
non_process
|
implement redesign and new components this is a meta ticket to track the ongoing redesign efforts in openverse this ticket is very much a work in progress and is currently being used to track missing ongoing issues with the redesign design todos general simplified homepage no featured collections yet redesigned content pages anything text heavy like sources and so on these shouldn t need much just a refresh with the latest colors sizing and so on mature content copyright claims add report this image audio flow to the single result pages img width alt cleanshot at src add back filter to enable mature content this can go in the filter sidebar the reporting modal needs to be updated in the new popover style img width alt cleanshot at src
| 0
|
11,553
| 14,435,280,100
|
IssuesEvent
|
2020-12-07 08:30:19
|
linuxdeepin/developer-center
|
https://api.github.com/repos/linuxdeepin/developer-center
|
closed
|
Control panel crash when click icons
|
bug | Ports bug | functional behavior other | delay processing suggest | functional behavior
|
Hi, i'm using deepin DE on Arch Linux, at the last upgrade when i switch from deepin-control-center-5.2.0.1 to deepin-control-center-5.2.0.3 i had a problem: the program crash when i select the account icon.
When i downgrade to the older version it works, i think that is a problem of the newest version.
|
1.0
|
Control panel crash when click icons - Hi, i'm using deepin DE on Arch Linux, at the last upgrade when i switch from deepin-control-center-5.2.0.1 to deepin-control-center-5.2.0.3 i had a problem: the program crash when i select the account icon.
When i downgrade to the older version it works, i think that is a problem of the newest version.
|
process
|
control panel crash when click icons hi i m using deepin de on arch linux at the last upgrade when i switch from deepin control center to deepin control center i had a problem the program crash when i select the account icon when i downgrade to the older version it works i think that is a problem of the newest version
| 1
|
23,709
| 12,086,734,667
|
IssuesEvent
|
2020-04-18 11:28:20
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
Dart high memory when open VS Code (macos)
|
dependency: dart framework platform-mac severe: performance tool
|
Hi, I'm getting slow when I open vs code that has installed Flutter plugins.
I look memory size of dart, consumed : 1,05 GB
Details :
```
cwd
/
txt
/Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/dart
txt
/Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/analysis_server.dart.snapshot
txt
/Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/kernel-service.dart.snapshot
txt
/usr/lib/dyld
txt
/private/var/db/dyld/dyld_shared_cache_x86_64
0
->0x6615e580cac70bb5
1
->0x6615e580cac71515
2
->0x6615e580cac707cd
3
->0x6615e580cac759a5
4
->0x6615e580cac75a65
5
count=0, state=0xa
6
->0x6615e580cac785e5
7
->0x6615e580cac75b25
10
->0x6615e580cac758e5
11
->0x6615e580cac78525
12
->0x6615e580cac786a5
13
->0x6615e580cac78465
37
/Applications/Visual Studio Code.app/Contents/Resources/app/node_modules.asar
```
and here the picture. Please fix it, if there are some bugs. Thank you.

|
True
|
Dart high memory when open VS Code (macos) - Hi, I'm getting slow when I open vs code that has installed Flutter plugins.
I look memory size of dart, consumed : 1,05 GB
Details :
```
cwd
/
txt
/Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/dart
txt
/Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/analysis_server.dart.snapshot
txt
/Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/kernel-service.dart.snapshot
txt
/usr/lib/dyld
txt
/private/var/db/dyld/dyld_shared_cache_x86_64
0
->0x6615e580cac70bb5
1
->0x6615e580cac71515
2
->0x6615e580cac707cd
3
->0x6615e580cac759a5
4
->0x6615e580cac75a65
5
count=0, state=0xa
6
->0x6615e580cac785e5
7
->0x6615e580cac75b25
10
->0x6615e580cac758e5
11
->0x6615e580cac78525
12
->0x6615e580cac786a5
13
->0x6615e580cac78465
37
/Applications/Visual Studio Code.app/Contents/Resources/app/node_modules.asar
```
and here the picture. Please fix it, if there are some bugs. Thank you.

|
non_process
|
dart high memory when open vs code macos hi i m getting slow when i open vs code that has installed flutter plugins i look memory size of dart consumed gb details cwd txt users macosx flutter bin cache dart sdk bin dart txt users macosx flutter bin cache dart sdk bin snapshots analysis server dart snapshot txt users macosx flutter bin cache dart sdk bin snapshots kernel service dart snapshot txt usr lib dyld txt private var db dyld dyld shared cache count state applications visual studio code app contents resources app node modules asar and here the picture please fix it if there are some bugs thank you
| 0
|
164,423
| 13,942,219,347
|
IssuesEvent
|
2020-10-22 20:40:04
|
TheRedDaemon/LittleCrusaderAsi
|
https://api.github.com/repos/TheRedDaemon/LittleCrusaderAsi
|
opened
|
Readme for implementation concepts still needs to be written.
|
documentation
|
It is already linkend and should contain something.
Otherwise it is not helpful...
|
1.0
|
Readme for implementation concepts still needs to be written. - It is already linkend and should contain something.
Otherwise it is not helpful...
|
non_process
|
readme for implementation concepts still needs to be written it is already linkend and should contain something otherwise it is not helpful
| 0
|
31,722
| 5,989,019,130
|
IssuesEvent
|
2017-06-02 07:18:11
|
icebob/vue-form-generator
|
https://api.github.com/repos/icebob/vue-form-generator
|
closed
|
update docs
|
difficulty: easy documentation
|
There are some recent changes what we need to update in documentation too.
- [x] `change` property in schema of file input #173
- [x] supported async validators #171
- [x] custom validation message for fields #169
- [x] supported string-based validators #167
- [x] vue-multiselect fixed #30 but in beta15, but they expanded [CSS from JS bundle](http://monterail.github.io/vue-multiselect/#sub-getting-started).
|
1.0
|
update docs - There are some recent changes what we need to update in documentation too.
- [x] `change` property in schema of file input #173
- [x] supported async validators #171
- [x] custom validation message for fields #169
- [x] supported string-based validators #167
- [x] vue-multiselect fixed #30 but in beta15, but they expanded [CSS from JS bundle](http://monterail.github.io/vue-multiselect/#sub-getting-started).
|
non_process
|
update docs there are some recent changes what we need to update in documentation too change property in schema of file input supported async validators custom validation message for fields supported string based validators vue multiselect fixed but in but they expanded
| 0
|
279,712
| 24,249,318,430
|
IssuesEvent
|
2022-09-27 13:08:16
|
hazelcast/hazelcast-python-client
|
https://api.github.com/repos/hazelcast/hazelcast-python-client
|
closed
|
test_translate_is_used
|
Type: Test-Failure Source: Internal
|
Failed on Windows against Python 2.7
https://github.com/hazelcast/hazelcast-python-client/runs/3889839024?check_suite_focus=true
```
======================================================================
FAIL: test_translate_is_used (tests.integration.connection_manager_translate_test.ConnectionManagerTranslateTest)
----------------------------------------------------------------------
Traceback (most recent call last):
File "D:\a\hazelcast-python-client\hazelcast-python-client\tests\integration\connection_manager_translate_test.py", line 51, in test_translate_is_used
connection_timeout=1.0,
AssertionError: IllegalStateError not raised
```
|
1.0
|
test_translate_is_used - Failed on Windows against Python 2.7
https://github.com/hazelcast/hazelcast-python-client/runs/3889839024?check_suite_focus=true
```
======================================================================
FAIL: test_translate_is_used (tests.integration.connection_manager_translate_test.ConnectionManagerTranslateTest)
----------------------------------------------------------------------
Traceback (most recent call last):
File "D:\a\hazelcast-python-client\hazelcast-python-client\tests\integration\connection_manager_translate_test.py", line 51, in test_translate_is_used
connection_timeout=1.0,
AssertionError: IllegalStateError not raised
```
|
non_process
|
test translate is used failed on windows against python fail test translate is used tests integration connection manager translate test connectionmanagertranslatetest traceback most recent call last file d a hazelcast python client hazelcast python client tests integration connection manager translate test py line in test translate is used connection timeout assertionerror illegalstateerror not raised
| 0
|
14,684
| 18,019,713,125
|
IssuesEvent
|
2021-09-16 17:47:08
|
lgblgblgb/xemu
|
https://api.github.com/repos/lgblgblgb/xemu
|
opened
|
MEGA65: V400 problem with non-bitplane modes?
|
target:MEGA65 compatibility-example VIC-IV
|
OpenROMs display only the top half of the screen, but not the bottom half, if 80x50 mode is used (it's OK 25 lines mode). Just trying to cursor down beyond the half of the screen, type something, and cursor up again: the cursor is again visible, the text is not, though in memory dump, the screen content clearly shows the typed text is there in the screen memory. According to FeralChild, 80x50 just works fine on the hardware.
My theory: OpenROMs may use V400 to achieve the 50 lines resolution which is not handled correctly (or at all) by Xemu. This is maybe true for "legacy" VIC-II modes (hires, MCM?) as well, I've never tested that (I even don't know if it's supported to have double vertical resolution with V400 set then). I'm quite sure, this is handled in bitplane mode though, since it was tested with various BASIC programs from the ClosedROMs project. So my conclusion: this is maybe not handled in other modes too much.
Note: `EFFICTIVE_V400` should be used instead of `REG_V400` macro because of some hacks which judges a certain situation as V200 even if it's V400.
|
True
|
MEGA65: V400 problem with non-bitplane modes? - OpenROMs display only the top half of the screen, but not the bottom half, if 80x50 mode is used (it's OK 25 lines mode). Just trying to cursor down beyond the half of the screen, type something, and cursor up again: the cursor is again visible, the text is not, though in memory dump, the screen content clearly shows the typed text is there in the screen memory. According to FeralChild, 80x50 just works fine on the hardware.
My theory: OpenROMs may use V400 to achieve the 50 lines resolution which is not handled correctly (or at all) by Xemu. This is maybe true for "legacy" VIC-II modes (hires, MCM?) as well, I've never tested that (I even don't know if it's supported to have double vertical resolution with V400 set then). I'm quite sure, this is handled in bitplane mode though, since it was tested with various BASIC programs from the ClosedROMs project. So my conclusion: this is maybe not handled in other modes too much.
Note: `EFFICTIVE_V400` should be used instead of `REG_V400` macro because of some hacks which judges a certain situation as V200 even if it's V400.
|
non_process
|
problem with non bitplane modes openroms display only the top half of the screen but not the bottom half if mode is used it s ok lines mode just trying to cursor down beyond the half of the screen type something and cursor up again the cursor is again visible the text is not though in memory dump the screen content clearly shows the typed text is there in the screen memory according to feralchild just works fine on the hardware my theory openroms may use to achieve the lines resolution which is not handled correctly or at all by xemu this is maybe true for legacy vic ii modes hires mcm as well i ve never tested that i even don t know if it s supported to have double vertical resolution with set then i m quite sure this is handled in bitplane mode though since it was tested with various basic programs from the closedroms project so my conclusion this is maybe not handled in other modes too much note effictive should be used instead of reg macro because of some hacks which judges a certain situation as even if it s
| 0
|
21,515
| 29,801,061,569
|
IssuesEvent
|
2023-06-16 08:08:43
|
openfoodfacts/openfoodfacts-server
|
https://api.github.com/repos/openfoodfacts/openfoodfacts-server
|
closed
|
The includes and their translations have not been deployed for Nova Groups pages
|
bug P1 static content nova processing Stale :star: top bug
|
### Describe the bug
The Nova Group pages seem to not be loading right, regardless of the amount of time I've tried, they aren't loading especially the fresh one, lots of information is missing.
### To Reproduce
https://fr.openfoodfacts.org/nova
https://world.openfoodfacts.org/nova
### Expected behavior
The pages should fully load and display the calculus methodology
|
1.0
|
The includes and their translations have not been deployed for Nova Groups pages - ### Describe the bug
The Nova Group pages seem to not be loading right, regardless of the amount of time I've tried, they aren't loading especially the fresh one, lots of information is missing.
### To Reproduce
https://fr.openfoodfacts.org/nova
https://world.openfoodfacts.org/nova
### Expected behavior
The pages should fully load and display the calculus methodology
|
process
|
the includes and their translations have not been deployed for nova groups pages describe the bug the nova group pages seem to not be loading right regardless of the amount of time i ve tried they aren t loading especially the fresh one lots of information is missing to reproduce expected behavior the pages should fully load and display the calculus methodology
| 1
|
21,825
| 30,316,774,062
|
IssuesEvent
|
2023-07-10 16:05:41
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
New Term - eventType
|
Term - add Class - Event normative Process - complete
|
## New term
* Submitter: John Wieczorek
* Efficacy Justification (why is this term necessary?): The hierarchical Event structure is not currently capable of distinguishing what different Events are for.
* Demand Justification (name at least two organizations that independently need this term): The need for an eventType to distinguish distinct kinds of activities has been promoted in the Interactions Interest Group (https://github.com/tdwg/interaction/issues/22), the Camera Trap Data Package repository (https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931487715 and https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931504444), in discussions about the vocabulary for basisOfRecord (https://github.com/tdwg/dwc/issues/302), in the design of conceptual and data publishing models in Diversifying the GBIF Data Model Project (https://tinyurl.com/diversifying-gbif-data), and for [OBIS](https://github.com/iobis/obis-issues/issues/172) .
* Stability Justification (what concerns are there that this might affect existing implementations?): The addition of this term will affect the definition of the GBIF Event Core, but this addition is already anticipated (see https://github.com/gbif/vocabulary/issues/107).
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: Yes, an equivalent term in the dwciri: namespace will be required.
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): eventType
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event
* Definition of the term (normative): The nature of the Event.
* Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use a controlled vocabulary. Regardless of the eventType, the interval of the Event should be captured in eventDate.
* Examples (not normative): `Sample`, `Observation`, `Site Visit`, `Biotic Interaction`, `Bioblitz`, `Expedition`, `Survey`, `Project` (There is a prospective vocabulary for consideration at https://registry.gbif-uat.org/vocabulary/EventType/concepts, but this proposal is not to define an eventType vocabulary).
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): Not in ABCD
|
1.0
|
New Term - eventType - ## New term
* Submitter: John Wieczorek
* Efficacy Justification (why is this term necessary?): The hierarchical Event structure is not currently capable of distinguishing what different Events are for.
* Demand Justification (name at least two organizations that independently need this term): The need for an eventType to distinguish distinct kinds of activities has been promoted in the Interactions Interest Group (https://github.com/tdwg/interaction/issues/22), the Camera Trap Data Package repository (https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931487715 and https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931504444), in discussions about the vocabulary for basisOfRecord (https://github.com/tdwg/dwc/issues/302), in the design of conceptual and data publishing models in Diversifying the GBIF Data Model Project (https://tinyurl.com/diversifying-gbif-data), and for [OBIS](https://github.com/iobis/obis-issues/issues/172) .
* Stability Justification (what concerns are there that this might affect existing implementations?): The addition of this term will affect the definition of the GBIF Event Core, but this addition is already anticipated (see https://github.com/gbif/vocabulary/issues/107).
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: Yes, an equivalent term in the dwciri: namespace will be required.
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): eventType
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event
* Definition of the term (normative): The nature of the Event.
* Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use a controlled vocabulary. Regardless of the eventType, the interval of the Event should be captured in eventDate.
* Examples (not normative): `Sample`, `Observation`, `Site Visit`, `Biotic Interaction`, `Bioblitz`, `Expedition`, `Survey`, `Project` (There is a prospective vocabulary for consideration at https://registry.gbif-uat.org/vocabulary/EventType/concepts, but this proposal is not to define an eventType vocabulary).
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): Not in ABCD
|
process
|
new term eventtype new term submitter john wieczorek efficacy justification why is this term necessary the hierarchical event structure is not currently capable of distinguishing what different events are for demand justification name at least two organizations that independently need this term the need for an eventtype to distinguish distinct kinds of activities has been promoted in the interactions interest group the camera trap data package repository and in discussions about the vocabulary for basisofrecord in the design of conceptual and data publishing models in diversifying the gbif data model project and for stability justification what concerns are there that this might affect existing implementations the addition of this term will affect the definition of the gbif event core but this addition is already anticipated see implications for dwciri namespace does this change affect a dwciri term version yes an equivalent term in the dwciri namespace will be required proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes eventtype organized in class e g occurrence event location taxon event definition of the term normative the nature of the event usage comments recommendations regarding content etc not normative recommended best practice is to use a controlled vocabulary regardless of the eventtype the interval of the event should be captured in eventdate examples not normative sample observation site visit biotic interaction bioblitz expedition survey project there is a prospective vocabulary for consideration at but this proposal is not to define an eventtype vocabulary refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
| 1
|
11,241
| 9,285,929,371
|
IssuesEvent
|
2019-03-21 08:58:14
|
astrolabsoftware/fink-broker
|
https://api.github.com/repos/astrolabsoftware/fink-broker
|
closed
|
Define common Apache Spark initialization step for services
|
apache kafka apache spark services
|
all services start with the same spark initialization steps (including setting the same Kafka parameters, etc). This is unnecessary redundancy, and prone to mistakes. The idea would be to factor this part, and include it into Fink core.
|
1.0
|
Define common Apache Spark initialization step for services - all services start with the same spark initialization steps (including setting the same Kafka parameters, etc). This is unnecessary redundancy, and prone to mistakes. The idea would be to factor this part, and include it into Fink core.
|
non_process
|
define common apache spark initialization step for services all services start with the same spark initialization steps including setting the same kafka parameters etc this is unnecessary redundancy and prone to mistakes the idea would be to factor this part and include it into fink core
| 0
|
166,064
| 14,018,951,902
|
IssuesEvent
|
2020-10-29 17:30:22
|
actions-pack/feedbacks
|
https://api.github.com/repos/actions-pack/feedbacks
|
closed
|
Compatibility with Advanced Custom Field Plugin?
|
documentation
|

You need to map your ACF field name in Registration Action's user meta fields. E.g one of the user meta field names is **phone** as shown in the image. So you just need to map your Elementor Form's Phone Field ID to **phone**
|
1.0
|
Compatibility with Advanced Custom Field Plugin? - 
You need to map your ACF field name in Registration Action's user meta fields. E.g one of the user meta field names is **phone** as shown in the image. So you just need to map your Elementor Form's Phone Field ID to **phone**
|
non_process
|
compatibility with advanced custom field plugin you need to map your acf field name in registration action s user meta fields e g one of the user meta field names is phone as shown in the image so you just need to map your elementor form s phone field id to phone
| 0
|
12,186
| 9,594,825,905
|
IssuesEvent
|
2019-05-09 14:46:31
|
opencb/opencga
|
https://api.github.com/repos/opencb/opencga
|
opened
|
Upload web service fails to upload to root folder
|
bug web services
|
files/upload web service fails when the user attempts to upload files to the root folder of the path.
|
1.0
|
Upload web service fails to upload to root folder - files/upload web service fails when the user attempts to upload files to the root folder of the path.
|
non_process
|
upload web service fails to upload to root folder files upload web service fails when the user attempts to upload files to the root folder of the path
| 0
|
36,750
| 6,548,340,108
|
IssuesEvent
|
2017-09-04 20:58:09
|
ekeih/OmNomNom
|
https://api.github.com/repos/ekeih/OmNomNom
|
opened
|
Improve documentation
|
documentation
|
Currently the bot is missing mostly everything you would expect from a well documented project.
If you are thinking about contributing code or just want to run the bot yourself, please let me know, so I can prioritize this issue ;-)
|
1.0
|
Improve documentation - Currently the bot is missing mostly everything you would expect from a well documented project.
If you are thinking about contributing code or just want to run the bot yourself, please let me know, so I can prioritize this issue ;-)
|
non_process
|
improve documentation currently the bot is missing mostly everything you would expect from a well documented project if you are thinking about contributing code or just want to run the bot yourself please let me know so i can prioritize this issue
| 0
|
17,338
| 23,157,876,125
|
IssuesEvent
|
2022-07-29 14:38:52
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[tabs] 希望支持选中的下标改色
|
good first issue in process
|
### 这个功能解决了什么问题

现在每次改色,都要手动去修改,极其不方便,希望能支持两种模式,纯改色,和slot的
### 你建议的方案是什么
把设置单独做成组件属性
|
1.0
|
[tabs] 希望支持选中的下标改色 - ### 这个功能解决了什么问题

现在每次改色,都要手动去修改,极其不方便,希望能支持两种模式,纯改色,和slot的
### 你建议的方案是什么
把设置单独做成组件属性
|
process
|
希望支持选中的下标改色 这个功能解决了什么问题 现在每次改色,都要手动去修改,极其不方便,希望能支持两种模式,纯改色,和slot的 你建议的方案是什么 把设置单独做成组件属性
| 1
|
5,335
| 8,154,440,234
|
IssuesEvent
|
2018-08-23 03:17:18
|
HumanCellAtlas/dcp-community
|
https://api.github.com/repos/HumanCellAtlas/dcp-community
|
closed
|
Objective or Objectives
|
charter-process
|
I think in the context of our charters Objective as found in https://github.com/HumanCellAtlas/dcp-community/blob/master/charters/charter-template.md tends to be used in the plural rather than singular
Can this be updated to the plural Objectives?
|
1.0
|
Objective or Objectives - I think in the context of our charters Objective as found in https://github.com/HumanCellAtlas/dcp-community/blob/master/charters/charter-template.md tends to be used in the plural rather than singular
Can this be updated to the plural Objectives?
|
process
|
objective or objectives i think in the context of our charters objective as found in tends to be used in the plural rather than singular can this be updated to the plural objectives
| 1
|
18,525
| 24,552,095,119
|
IssuesEvent
|
2022-10-12 13:21:37
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] [Offline indicator] Offline feature is not working in the following scenario
|
Bug P1 iOS Process: Fixed Process: Tested dev
|
**Description**
**Pre_condition:** The testing device should connect to WiFi
**Steps:**
1. Install the mobile app on the testing device
2. Turn off the main data (Don't turn off the WI-FI of the testing device)
3. Observe
**AR:** Offline indicator feature is not working
**ER:** Offline indicator feature should work as per expected in the above scenario
Note: Issue is not observed in android
|
2.0
|
[iOS] [Offline indicator] Offline feature is not working in the following scenario - **Description**
**Pre_condition:** The testing device should connect to WiFi
**Steps:**
1. Install the mobile app on the testing device
2. Turn off the main data (Don't turn off the WI-FI of the testing device)
3. Observe
**AR:** Offline indicator feature is not working
**ER:** Offline indicator feature should work as per expected in the above scenario
Note: Issue is not observed in android
|
process
|
offline feature is not working in the following scenario description pre condition the testing device should connect to wifi steps install the mobile app on the testing device turn off the main data don t turn off the wi fi of the testing device observe ar offline indicator feature is not working er offline indicator feature should work as per expected in the above scenario note issue is not observed in android
| 1
|
1,703
| 4,349,928,413
|
IssuesEvent
|
2016-07-30 22:27:31
|
pwittchen/ReactiveSensors
|
https://api.github.com/repos/pwittchen/ReactiveSensors
|
closed
|
Release 0.1.2
|
release process
|
**Initial release notes**:
- bumped RxJava dependency to v. 1.1.8
- bumped RxAndroid dependency to v. 1.2.1
- bumped Google Truth test dependency to v. 0.28
- bumped Compile SDK version to v. 23
- bumped Kotlin to v. 1.0.0
- updated sample apps
**Things to do**:
- [x] bump library version
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] create new GitHub release
|
1.0
|
Release 0.1.2 - **Initial release notes**:
- bumped RxJava dependency to v. 1.1.8
- bumped RxAndroid dependency to v. 1.2.1
- bumped Google Truth test dependency to v. 0.28
- bumped Compile SDK version to v. 23
- bumped Kotlin to v. 1.0.0
- updated sample apps
**Things to do**:
- [x] bump library version
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] create new GitHub release
|
process
|
release initial release notes bumped rxjava dependency to v bumped rxandroid dependency to v bumped google truth test dependency to v bumped compile sdk version to v bumped kotlin to v updated sample apps things to do bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
| 1
|
427,697
| 12,397,948,738
|
IssuesEvent
|
2020-05-21 00:15:28
|
eclipse-ee4j/glassfish
|
https://api.github.com/repos/eclipse-ee4j/glassfish
|
closed
|
[osgi-cdi]OSGi service automatic publishing with @Publish-liking annotation
|
Component: OSGi-JavaEE ERR: Assignee Priority: Critical Stale Type: New Feature
|
Liking Weld-OSGi:
allows developers to automatically publish service implementation.There is nothing to do, just put the
annotation. OSGi framework is completely hidden. Then the service is accessible through CDI-OSGi service injection
and OSGi classic mechanisms.
In addition, on OSGi RFP-0146 Draft,
CDI002 - The specification MUST make it possible to publish CDI beans in the OSGi Service Registry.
So, this is a critical Requirement on CDI/OSGi Integration just as @OSGiService.
#### Affected Versions
[future release]
|
1.0
|
[osgi-cdi]OSGi service automatic publishing with @Publish-liking annotation - Liking Weld-OSGi:
allows developers to automatically publish service implementation.There is nothing to do, just put the
annotation. OSGi framework is completely hidden. Then the service is accessible through CDI-OSGi service injection
and OSGi classic mechanisms.
In addition, on OSGi RFP-0146 Draft,
CDI002 - The specification MUST make it possible to publish CDI beans in the OSGi Service Registry.
So, this is a critical Requirement on CDI/OSGi Integration just as @OSGiService.
#### Affected Versions
[future release]
|
non_process
|
osgi service automatic publishing with publish liking annotation liking weld osgi allows developers to automatically publish service implementation there is nothing to do just put the annotation osgi framework is completely hidden then the service is accessible through cdi osgi service injection and osgi classic mechanisms in addition on osgi rfp draft the specification must make it possible to publish cdi beans in the osgi service registry so this is a critical requirement on cdi osgi integration just as osgiservice affected versions
| 0
|
19,070
| 25,098,729,614
|
IssuesEvent
|
2022-11-08 12:07:55
|
hoprnet/hoprnet
|
https://api.github.com/repos/hoprnet/hoprnet
|
closed
|
Add staging branches for all supported releases
|
devops epic processes
|
We want to be able to merge PRs into release staging branches and only when we consider the sum of changes release-worthy, that branch is merged into the release branch. This process should be started with #4275
# Example
Release branch: `release/bogota`
Staging branch: `release-staging/bogota`
1. PRs are merged into `release-staging/bogota`
2. CI deployment workflow runs through and create Docker images and deploys nodes.
2.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2-next.3`
2.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota-staging`
3. Merge PR into branch `release/bogota` is created and merged
4. CI deployment workflow runs through and create Docker images and deploys nodes.
4.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2`
4.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota`
# Requirements
- [x] CI workflow is adapted to support execution on staging branches
- [x] protocol config is adapted to support per release staging versions in the same environment
- [x] version generation script, see `scripts/publish-version.sh` can generated staging release versions
|
1.0
|
Add staging branches for all supported releases - We want to be able to merge PRs into release staging branches and only when we consider the sum of changes release-worthy, that branch is merged into the release branch. This process should be started with #4275
# Example
Release branch: `release/bogota`
Staging branch: `release-staging/bogota`
1. PRs are merged into `release-staging/bogota`
2. CI deployment workflow runs through and create Docker images and deploys nodes.
2.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2-next.3`
2.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota-staging`
3. Merge PR into branch `release/bogota` is created and merged
4. CI deployment workflow runs through and create Docker images and deploys nodes.
4.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2`
4.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota`
# Requirements
- [x] CI workflow is adapted to support execution on staging branches
- [x] protocol config is adapted to support per release staging versions in the same environment
- [x] version generation script, see `scripts/publish-version.sh` can generated staging release versions
|
process
|
add staging branches for all supported releases we want to be able to merge prs into release staging branches and only when we consider the sum of changes release worthy that branch is merged into the release branch this process should be started with example release branch release bogota staging branch release staging bogota prs are merged into release staging bogota ci deployment workflow runs through and create docker images and deploys nodes docker image name gcr io hoprassociation hoprd next moving docker tag gcr io hoprassociation hoprd bogota staging merge pr into branch release bogota is created and merged ci deployment workflow runs through and create docker images and deploys nodes docker image name gcr io hoprassociation hoprd moving docker tag gcr io hoprassociation hoprd bogota requirements ci workflow is adapted to support execution on staging branches protocol config is adapted to support per release staging versions in the same environment version generation script see scripts publish version sh can generated staging release versions
| 1
|
2,598
| 5,356,200,627
|
IssuesEvent
|
2017-02-20 15:04:35
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Bootstrap Table Warnings
|
enhancement inprocess
|
Hi Allen,
Thanks again for taking your time everyday to read through these issues. It is much appreciated. I'm getting the following warnings when I render your tables:

This doesn't break the tables as of now, but is obviously a concern for the future.
Currently using version 2.11.0.
|
1.0
|
Bootstrap Table Warnings - Hi Allen,
Thanks again for taking your time everyday to read through these issues. It is much appreciated. I'm getting the following warnings when I render your tables:

This doesn't break the tables as of now, but is obviously a concern for the future.
Currently using version 2.11.0.
|
process
|
bootstrap table warnings hi allen thanks again for taking your time everyday to read through these issues it is much appreciated i m getting the following warnings when i render your tables this doesn t break the tables as of now but is obviously a concern for the future currently using version
| 1
|
57,767
| 24,223,083,146
|
IssuesEvent
|
2022-09-26 12:30:34
|
adi-H/top-picks
|
https://api.github.com/repos/adi-H/top-picks
|
closed
|
add change review data to product page
|
enhancement ui list-microservice ratings-microservice to be continued backend
|
continution of TBD
**updated 27/8**
ratings
- [x] edit / update rating route + tests
lists
- [ ] find product in user lists
ui
- [ ] add ur own review
- [ ] change ur review
|
2.0
|
add change review data to product page - continution of TBD
**updated 27/8**
ratings
- [x] edit / update rating route + tests
lists
- [ ] find product in user lists
ui
- [ ] add ur own review
- [ ] change ur review
|
non_process
|
add change review data to product page continution of tbd updated ratings edit update rating route tests lists find product in user lists ui add ur own review change ur review
| 0
|
2,670
| 5,468,638,854
|
IssuesEvent
|
2017-03-10 07:04:26
|
openslide/openslide
|
https://api.github.com/repos/openslide/openslide
|
closed
|
Audit security options pages
|
development-process enhancement
|
There are lots of new options for teams and collaboration now. We should go through and decide if any changes are needed.
https://github.com/organizations/openslide/settings/member_privileges
https://github.com/organizations/openslide/settings/oauth_application_policy
https://github.com/openslide/openslide/settings/collaboration
|
1.0
|
Audit security options pages - There are lots of new options for teams and collaboration now. We should go through and decide if any changes are needed.
https://github.com/organizations/openslide/settings/member_privileges
https://github.com/organizations/openslide/settings/oauth_application_policy
https://github.com/openslide/openslide/settings/collaboration
|
process
|
audit security options pages there are lots of new options for teams and collaboration now we should go through and decide if any changes are needed
| 1
|
337,745
| 30,259,979,870
|
IssuesEvent
|
2023-07-07 07:26:32
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: failover/non-system/pause failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker T-kv
|
roachtest.failover/non-system/pause [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=artifacts#/failover/non-system/pause) on master @ [dbe8511fae8fca21562fdde5c240b1f7d06ef582](https://github.com/cockroachdb/cockroach/commits/dbe8511fae8fca21562fdde5c240b1f7d06ef582):
```
(test_runner.go:1075).runTest: test timed out (30m0s)
(cluster.go:2282).Run: output in run_141100.736540940_n7_cockroach-workload-i: ./cockroach workload init kv --splits 1000 {pgurl:1} returned: COMMAND_PROBLEM: exit status 137
test artifacts and logs in: /artifacts/failover/non-system/pause/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/non-system/pause.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29488
|
2.0
|
roachtest: failover/non-system/pause failed - roachtest.failover/non-system/pause [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=artifacts#/failover/non-system/pause) on master @ [dbe8511fae8fca21562fdde5c240b1f7d06ef582](https://github.com/cockroachdb/cockroach/commits/dbe8511fae8fca21562fdde5c240b1f7d06ef582):
```
(test_runner.go:1075).runTest: test timed out (30m0s)
(cluster.go:2282).Run: output in run_141100.736540940_n7_cockroach-workload-i: ./cockroach workload init kv --splits 1000 {pgurl:1} returned: COMMAND_PROBLEM: exit status 137
test artifacts and logs in: /artifacts/failover/non-system/pause/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/non-system/pause.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29488
|
non_process
|
roachtest failover non system pause failed roachtest failover non system pause with on master test runner go runtest test timed out cluster go run output in run cockroach workload i cockroach workload init kv splits pgurl returned command problem exit status test artifacts and logs in artifacts failover non system pause run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
| 0
|
473,818
| 13,648,059,297
|
IssuesEvent
|
2020-09-26 07:03:44
|
TerryCavanagh/diceydungeonsbeta
|
https://api.github.com/repos/TerryCavanagh/diceydungeonsbeta
|
closed
|
Precious Egg and similar should be immune to fury/have unique behavior
|
High Priority release candidate v0.1
|
When furied, precious egg duplicates, and this can make your skillcard inaccessible which is potentially run-breaking. Remix robot getting fury is really common so this can be a problem! Maybe it should just do nothing when used with fury, or skip a step? (Ex. precious egg 6 skips to 4, precious egg 2 just gives you the item inside)
|
1.0
|
Precious Egg and similar should be immune to fury/have unique behavior - When furied, precious egg duplicates, and this can make your skillcard inaccessible which is potentially run-breaking. Remix robot getting fury is really common so this can be a problem! Maybe it should just do nothing when used with fury, or skip a step? (Ex. precious egg 6 skips to 4, precious egg 2 just gives you the item inside)
|
non_process
|
precious egg and similar should be immune to fury have unique behavior when furied precious egg duplicates and this can make your skillcard inaccessible which is potentially run breaking remix robot getting fury is really common so this can be a problem maybe it should just do nothing when used with fury or skip a step ex precious egg skips to precious egg just gives you the item inside
| 0
|
2,007
| 4,827,340,176
|
IssuesEvent
|
2016-11-07 13:19:24
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
closed
|
Fix identifiers extraction in EUCTR
|
3. In Development bug Processors
|
Currently we just [prepend "EUCTR" to `eudract_number`](https://github.com/opentrials/processors/blob/master/processors/euctr/extractors.py#L27) but there are [records whose `eudract_number` looks like `EUCTR2014-001259-22-3rd`](https://query.opentrials.net/queries/140). They will end up [having no identifiers](https://explorer.opentrials.net/trials/42a1b0a4-8c52-11e6-a776-0242ac12000b).
We should prepend smartly to avoid turning identifiers like the one above into invalid ones (`EUCTREUCTR2014-001259-22-3rd`).
`euctr` processor is the one that has been observed but there may be others with this issue so they should all be checked.
|
1.0
|
Fix identifiers extraction in EUCTR - Currently we just [prepend "EUCTR" to `eudract_number`](https://github.com/opentrials/processors/blob/master/processors/euctr/extractors.py#L27) but there are [records whose `eudract_number` looks like `EUCTR2014-001259-22-3rd`](https://query.opentrials.net/queries/140). They will end up [having no identifiers](https://explorer.opentrials.net/trials/42a1b0a4-8c52-11e6-a776-0242ac12000b).
We should prepend smartly to avoid turning identifiers like the one above into invalid ones (`EUCTREUCTR2014-001259-22-3rd`).
`euctr` processor is the one that has been observed but there may be others with this issue so they should all be checked.
|
process
|
fix identifiers extraction in euctr currently we just but there are they will end up we should prepend smartly to avoid turning identifiers like the one above into invalid ones euctr processor is the one that has been observed but there may be others with this issue so they should all be checked
| 1
|
8,540
| 11,714,073,216
|
IssuesEvent
|
2020-03-09 11:34:44
|
kazuwjnlab/cvpaper
|
https://api.github.com/repos/kazuwjnlab/cvpaper
|
opened
|
[cvpaper] CVPR2019 #686 Libra R-CNN: Towards Balanced Learning for Object Detection
|
Object Detection imbalance training process
|
## \#686 [Libra R-CNN: Towards Balanced Learning for Object Detection](http://openaccess.thecvf.com/content_CVPR_2019/papers/Pang_Libra_R-CNN_Towards_Balanced_Learning_for_Object_Detection_CVPR_2019_paper.pdf)
Jiangmiao Pang, Kai Chen, Jianping Shi, Huajun Feng, Wanli Ouyang, Dahua Lin
### どんな論文か?
この研究は、物体検出の標準的な訓練方法を再検討します。 この論文では、検出性能についての言及は、トレーニングプロセス中の3つのレベルからなる不均衡によってしばしば制限されています。 1)サンプルレベル、2)特徴レベル、および3)客観的レベル。 この観察の結果として、本論文は天秤座R-CNN、物体検出のためのバランスのとれた学習に向けた簡単だが効果的なフレームワークを提案する。
### 新規性
3つの要素を組み合わせることによってトレーニングプロセスのバランスを取り直すフレームワーク:a)IoUバランスサンプリング、追加費用なしのシンプルだが効果的なハードマイニング方法、b)同じ深く統合されたバランスのとれた意味機能を使用したマルチレベル機能の強化 、c)関連するサンプルおよびタスクをバランスを取り戻すための重大な回帰勾配を促進するための、L1損失の均衡化
### 結果
この手法をMS COCOデータセットで評価し、最先端のオブジェクト検出手法と比較しました。 全体的にバランスのとれた設計により、提案されたLibra R-CNNは38.7 APを達成し、これはFPN Faster R-CNNよりも2.5ポイント高くなっています。 これは、はるかに強力な機能抽出機能であるResNetXt-101-64x4dで後者よりも優れています。
### その他(なぜ通ったか?等)
著者らは彼らの研究を単一の検出器、Libra RetinaNet(Libra RN)に拡張した。 T.天秤座RPNからの大幅な改善。 結果を添付の表3に示す。

object detection, training process, imbalance,
|
1.0
|
[cvpaper] CVPR2019 #686 Libra R-CNN: Towards Balanced Learning for Object Detection - ## \#686 [Libra R-CNN: Towards Balanced Learning for Object Detection](http://openaccess.thecvf.com/content_CVPR_2019/papers/Pang_Libra_R-CNN_Towards_Balanced_Learning_for_Object_Detection_CVPR_2019_paper.pdf)
Jiangmiao Pang, Kai Chen, Jianping Shi, Huajun Feng, Wanli Ouyang, Dahua Lin
### どんな論文か?
この研究は、物体検出の標準的な訓練方法を再検討します。 この論文では、検出性能についての言及は、トレーニングプロセス中の3つのレベルからなる不均衡によってしばしば制限されています。 1)サンプルレベル、2)特徴レベル、および3)客観的レベル。 この観察の結果として、本論文は天秤座R-CNN、物体検出のためのバランスのとれた学習に向けた簡単だが効果的なフレームワークを提案する。
### 新規性
3つの要素を組み合わせることによってトレーニングプロセスのバランスを取り直すフレームワーク:a)IoUバランスサンプリング、追加費用なしのシンプルだが効果的なハードマイニング方法、b)同じ深く統合されたバランスのとれた意味機能を使用したマルチレベル機能の強化 、c)関連するサンプルおよびタスクをバランスを取り戻すための重大な回帰勾配を促進するための、L1損失の均衡化
### 結果
この手法をMS COCOデータセットで評価し、最先端のオブジェクト検出手法と比較しました。 全体的にバランスのとれた設計により、提案されたLibra R-CNNは38.7 APを達成し、これはFPN Faster R-CNNよりも2.5ポイント高くなっています。 これは、はるかに強力な機能抽出機能であるResNetXt-101-64x4dで後者よりも優れています。
### その他(なぜ通ったか?等)
著者らは彼らの研究を単一の検出器、Libra RetinaNet(Libra RN)に拡張した。 T.天秤座RPNからの大幅な改善。 結果を添付の表3に示す。

object detection, training process, imbalance,
|
process
|
libra r cnn towards balanced learning for object detection jiangmiao pang kai chen jianping shi huajun feng wanli ouyang dahua lin どんな論文か? この研究は、物体検出の標準的な訓練方法を再検討します。 この論文では、検出性能についての言及は、 。 )サンプルレベル、 )特徴レベル、 )客観的レベル。 この観察の結果として、本論文は天秤座r cnn、物体検出のためのバランスのとれた学習に向けた簡単だが効果的なフレームワークを提案する。 新規性 :a)iouバランスサンプリング、追加費用なしのシンプルだが効果的なハードマイニング方法、b)同じ深く統合されたバランスのとれた意味機能を使用したマルチレベル機能の強化 、c)関連するサンプルおよびタスクをバランスを取り戻すための重大な回帰勾配を促進するための、 結果 この手法をms cocoデータセットで評価し、最先端のオブジェクト検出手法と比較しました。 全体的にバランスのとれた設計により、提案されたlibra r apを達成し、これはfpn faster r 。 これは、はるかに強力な機能抽出機能であるresnetxt 。 その他(なぜ通ったか?等) 著者らは彼らの研究を単一の検出器、libra retinanet(libra rn)に拡張した。 t 天秤座rpnからの大幅な改善。 。 object detection training process imbalance
| 1
|
21,780
| 30,294,577,725
|
IssuesEvent
|
2023-07-09 17:44:29
|
The-Data-Alchemists-Manipal/MindWave
|
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
|
closed
|
Eye Detection Using opencv
|
image-processing
|
## 💥 Proposal
In this project, I am going to use open cv to detect eyes in the image/video.
|
1.0
|
Eye Detection Using opencv - ## 💥 Proposal
In this project, I am going to use open cv to detect eyes in the image/video.
|
process
|
eye detection using opencv 💥 proposal in this project i am going to use open cv to detect eyes in the image video
| 1
|
57,139
| 8,142,320,778
|
IssuesEvent
|
2018-08-21 07:10:16
|
github/orchestrator
|
https://api.github.com/repos/github/orchestrator
|
closed
|
Differentiating GracefulTakeover from MasterFailure in Hook
|
documentation
|
I noticed that the Recovery Hooks do not seem to differentiate between planned failovers (Graceful-Master-Takeover) and actual failures.
I had been having an issue where I has triggering a STONITH process in my PreFailoverProcesses and noticed that this was making my GracefulTakeovers fail. What was happening was I did not expect the Graceful Takeover to trigger the PreFailoverProcesses as there were seperate hooks for these so I assumed wrong. Thus, my STONITH process was triggering mid-takeover and shutting down the master.
This is fine, though I think the documentation could better illuminate what each hook is for, except that I'm not actually sure how to differentiate the difference between planned takeovers and true failures. In a graceful takeover, I don't really want to STONITH my master as I trust the takeover process to handle the rotation (or maybe this is a bad assumption on my part?).
I thought maybe ORC_FAILURE_TYPE ({failureType}) would work, but it indicates "DeadMaster". I checked the source to find the list of possible values and noticed there is a "NoProblem" state that doesn't appear to be used. Why does Orchestrator label a Takeover as "DeadMaster". I would expect something like "PlannedSwitch" or "Takeover" as my Master isn't really dead.
Anyways, I tried looking in the source that generates the magic replacement variables and the environment variables and noticed one that is not documented in the md files: ORC_COMMAND. I added this to my hooks and viola: "ORC_COMMAND: graceful-master-takeover". Excellent! I could pass this value to my STONITH process and have it either run a system check or do nothing. However, since it's undocumented, i'm not sure if it's wise to use it. Is this just a missing documentation item, or should I not be using this variable in this manner?
Thanks
|
1.0
|
Differentiating GracefulTakeover from MasterFailure in Hook - I noticed that the Recovery Hooks do not seem to differentiate between planned failovers (Graceful-Master-Takeover) and actual failures.
I had been having an issue where I has triggering a STONITH process in my PreFailoverProcesses and noticed that this was making my GracefulTakeovers fail. What was happening was I did not expect the Graceful Takeover to trigger the PreFailoverProcesses as there were seperate hooks for these so I assumed wrong. Thus, my STONITH process was triggering mid-takeover and shutting down the master.
This is fine, though I think the documentation could better illuminate what each hook is for, except that I'm not actually sure how to differentiate the difference between planned takeovers and true failures. In a graceful takeover, I don't really want to STONITH my master as I trust the takeover process to handle the rotation (or maybe this is a bad assumption on my part?).
I thought maybe ORC_FAILURE_TYPE ({failureType}) would work, but it indicates "DeadMaster". I checked the source to find the list of possible values and noticed there is a "NoProblem" state that doesn't appear to be used. Why does Orchestrator label a Takeover as "DeadMaster". I would expect something like "PlannedSwitch" or "Takeover" as my Master isn't really dead.
Anyways, I tried looking in the source that generates the magic replacement variables and the environment variables and noticed one that is not documented in the md files: ORC_COMMAND. I added this to my hooks and viola: "ORC_COMMAND: graceful-master-takeover". Excellent! I could pass this value to my STONITH process and have it either run a system check or do nothing. However, since it's undocumented, i'm not sure if it's wise to use it. Is this just a missing documentation item, or should I not be using this variable in this manner?
Thanks
|
non_process
|
differentiating gracefultakeover from masterfailure in hook i noticed that the recovery hooks do not seem to differentiate between planned failovers graceful master takeover and actual failures i had been having an issue where i has triggering a stonith process in my prefailoverprocesses and noticed that this was making my gracefultakeovers fail what was happening was i did not expect the graceful takeover to trigger the prefailoverprocesses as there were seperate hooks for these so i assumed wrong thus my stonith process was triggering mid takeover and shutting down the master this is fine though i think the documentation could better illuminate what each hook is for except that i m not actually sure how to differentiate the difference between planned takeovers and true failures in a graceful takeover i don t really want to stonith my master as i trust the takeover process to handle the rotation or maybe this is a bad assumption on my part i thought maybe orc failure type failuretype would work but it indicates deadmaster i checked the source to find the list of possible values and noticed there is a noproblem state that doesn t appear to be used why does orchestrator label a takeover as deadmaster i would expect something like plannedswitch or takeover as my master isn t really dead anyways i tried looking in the source that generates the magic replacement variables and the environment variables and noticed one that is not documented in the md files orc command i added this to my hooks and viola orc command graceful master takeover excellent i could pass this value to my stonith process and have it either run a system check or do nothing however since it s undocumented i m not sure if it s wise to use it is this just a missing documentation item or should i not be using this variable in this manner thanks
| 0
|
7,229
| 10,368,257,304
|
IssuesEvent
|
2019-09-07 15:29:55
|
banctilrobitaille/kerosene
|
https://api.github.com/repos/banctilrobitaille/kerosene
|
closed
|
[FEATURE] Implement a standard api for custom variable ploting
|
EventPreprocessor ploting
|
**Is your feature request related to a problem? Please describe.**
No.
**Describe the solution you'd like**
Instead of having different EventPreprocessor for every visdom plot type we should have a single event preprocessor for custom variables that accepts the type plot, frequency and opts as args.
**Describe alternatives you've considered**
none
**Will this change the current api? How?**
Yes, instead of having multiple Preprocessor for custom variable there will be one.
**Additional context**
No.
|
1.0
|
[FEATURE] Implement a standard api for custom variable ploting - **Is your feature request related to a problem? Please describe.**
No.
**Describe the solution you'd like**
Instead of having different EventPreprocessor for every visdom plot type we should have a single event preprocessor for custom variables that accepts the type plot, frequency and opts as args.
**Describe alternatives you've considered**
none
**Will this change the current api? How?**
Yes, instead of having multiple Preprocessor for custom variable there will be one.
**Additional context**
No.
|
process
|
implement a standard api for custom variable ploting is your feature request related to a problem please describe no describe the solution you d like instead of having different eventpreprocessor for every visdom plot type we should have a single event preprocessor for custom variables that accepts the type plot frequency and opts as args describe alternatives you ve considered none will this change the current api how yes instead of having multiple preprocessor for custom variable there will be one additional context no
| 1
|
69,926
| 30,505,008,875
|
IssuesEvent
|
2023-07-18 16:13:59
|
hashicorp/terraform-provider-azurerm
|
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
|
closed
|
Error: could not find DNS zone
|
bug service/dns
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.9
### AzureRM Provider Version
3.46.0
### Affected Resource(s)/Data Source(s)
azurerm_dns_zone
### Terraform Configuration Files
```hcl
data "azurerm_dns_zone" "this" {
name = var.domain.name
}
```
### Debug Output/Panic Output
```shell
╷
│ Error: could not find DNS zone with name: "foo.com"
│
│ with module.cdn[0].data.azurerm_dns_zone.this[0],
│ on modules/cdn/cdn.tf line 44, in data "azurerm_dns_zone" "this":
│ 44: data "azurerm_dns_zone" "this" {
│
╵
```
### Expected Behaviour
Work
### Actual Behaviour
Doesn't work.
### Steps to Reproduce
_No response_
### Important Factoids
It works with hashicorp/azurerm = 3.45.0 and earlier.
### References
_No response_
|
1.0
|
Error: could not find DNS zone - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.9
### AzureRM Provider Version
3.46.0
### Affected Resource(s)/Data Source(s)
azurerm_dns_zone
### Terraform Configuration Files
```hcl
data "azurerm_dns_zone" "this" {
name = var.domain.name
}
```
### Debug Output/Panic Output
```shell
╷
│ Error: could not find DNS zone with name: "foo.com"
│
│ with module.cdn[0].data.azurerm_dns_zone.this[0],
│ on modules/cdn/cdn.tf line 44, in data "azurerm_dns_zone" "this":
│ 44: data "azurerm_dns_zone" "this" {
│
╵
```
### Expected Behaviour
Work
### Actual Behaviour
Doesn't work.
### Steps to Reproduce
_No response_
### Important Factoids
It works with hashicorp/azurerm = 3.45.0 and earlier.
### References
_No response_
|
non_process
|
error could not find dns zone is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s azurerm dns zone terraform configuration files hcl data azurerm dns zone this name var domain name debug output panic output shell ╷ │ error could not find dns zone with name foo com │ │ with module cdn data azurerm dns zone this │ on modules cdn cdn tf line in data azurerm dns zone this │ data azurerm dns zone this │ ╵ expected behaviour work actual behaviour doesn t work steps to reproduce no response important factoids it works with hashicorp azurerm and earlier references no response
| 0
|
9,944
| 11,948,421,548
|
IssuesEvent
|
2020-04-03 11:49:57
|
icatproject/python-icat
|
https://api.github.com/repos/icatproject/python-icat
|
opened
|
Consider to switch to suds-community
|
compatibility
|
The original version of `suds` is dead since very long time. Jurko Gospodnetić has been taken over and created the `suds-jurko` fork. But development seem to have stalled again since 2015. Many new forks have been created since then, most of them rather short-lived. Now, there is one that lasted a little bit longer and at least has recent activity: [suds-community](https://pypi.org/project/suds-community/).
Should test the latter and evaluate whether it works with python-icat. If positive, update documentation to point to that one as system requirement in the README.
|
True
|
Consider to switch to suds-community - The original version of `suds` is dead since very long time. Jurko Gospodnetić has been taken over and created the `suds-jurko` fork. But development seem to have stalled again since 2015. Many new forks have been created since then, most of them rather short-lived. Now, there is one that lasted a little bit longer and at least has recent activity: [suds-community](https://pypi.org/project/suds-community/).
Should test the latter and evaluate whether it works with python-icat. If positive, update documentation to point to that one as system requirement in the README.
|
non_process
|
consider to switch to suds community the original version of suds is dead since very long time jurko gospodnetić has been taken over and created the suds jurko fork but development seem to have stalled again since many new forks have been created since then most of them rather short lived now there is one that lasted a little bit longer and at least has recent activity should test the latter and evaluate whether it works with python icat if positive update documentation to point to that one as system requirement in the readme
| 0
|
729,081
| 25,108,544,443
|
IssuesEvent
|
2022-11-08 18:30:10
|
cse-sim/cse
|
https://api.github.com/repos/cse-sim/cse
|
opened
|
Review and consider localizing environment-specific code
|
1 - low priority
|
The original idea of envpak was to isolate OS-specific and/or compiler-specific code -- provide a uniform internal API for the associated features.
There are OS and compiler dependencies cropping up in other files. Notably rmkerr, but others also.
Those dependencies could be reviewed and some capabilities moved to envpak as sensible / practical.
|
1.0
|
Review and consider localizing environment-specific code - The original idea of envpak was to isolate OS-specific and/or compiler-specific code -- provide a uniform internal API for the associated features.
There are OS and compiler dependencies cropping up in other files. Notably rmkerr, but others also.
Those dependencies could be reviewed and some capabilities moved to envpak as sensible / practical.
|
non_process
|
review and consider localizing environment specific code the original idea of envpak was to isolate os specific and or compiler specific code provide a uniform internal api for the associated features there are os and compiler dependencies cropping up in other files notably rmkerr but others also those dependencies could be reviewed and some capabilities moved to envpak as sensible practical
| 0
|
611,590
| 18,959,053,651
|
IssuesEvent
|
2021-11-19 00:51:38
|
GoogleChrome/lighthouse
|
https://api.github.com/repos/GoogleChrome/lighthouse
|
closed
|
Lighthouse test feature in Chrome just stopped working
|
bug needs-priority
|
### FAQ
- [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/master/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/master/docs/throttling.md).
- [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead).
### URL
https://m.warsztat24.pl
### What happened?
I cannot perform lighthouse test on https://m.warsztat24.pl for a few days now... there wasn't any problems before... but now... test runs and runs and runs (messages on popup do change all over the time) and just cannot finish and go to the summary....
what can be the reason? please help!
### What did you expect?
I would like just to get lighthouse test results for https://m.warsztat24.pl.
### What have you tried?
don't know what could I do, as the tests seem continue over and over...
### How were you running Lighthouse?
Chrome DevTools
### Lighthouse Version
don't know because I cannot see any finished report
### Chrome Version
94.0.4606.81
### Node Version
12.16.1
### Relevant log output
_No response_
|
1.0
|
Lighthouse test feature in Chrome just stopped working - ### FAQ
- [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/master/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/master/docs/throttling.md).
- [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead).
### URL
https://m.warsztat24.pl
### What happened?
I cannot perform lighthouse test on https://m.warsztat24.pl for a few days now... there wasn't any problems before... but now... test runs and runs and runs (messages on popup do change all over the time) and just cannot finish and go to the summary....
what can be the reason? please help!
### What did you expect?
I would like just to get lighthouse test results for https://m.warsztat24.pl.
### What have you tried?
don't know what could I do, as the tests seem continue over and over...
### How were you running Lighthouse?
Chrome DevTools
### Lighthouse Version
don't know because I cannot see any finished report
### Chrome Version
94.0.4606.81
### Node Version
12.16.1
### Relevant log output
_No response_
|
non_process
|
lighthouse test feature in chrome just stopped working faq yes my issue is not about or yes my issue is not about a specific accessibility audit file with instead url what happened i cannot perform lighthouse test on for a few days now there wasn t any problems before but now test runs and runs and runs messages on popup do change all over the time and just cannot finish and go to the summary what can be the reason please help what did you expect i would like just to get lighthouse test results for what have you tried don t know what could i do as the tests seem continue over and over how were you running lighthouse chrome devtools lighthouse version don t know because i cannot see any finished report chrome version node version relevant log output no response
| 0
|
10,147
| 13,044,162,542
|
IssuesEvent
|
2020-07-29 03:47:33
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `Sleep` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `Sleep` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `Sleep` from TiDB -
## Description
Port the scalar function `Sleep` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function sleep from tidb description port the scalar function sleep from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
372,775
| 26,019,138,222
|
IssuesEvent
|
2022-12-21 11:02:37
|
boostercloud/booster
|
https://api.github.com/repos/boostercloud/booster
|
opened
|
Write documentation on how to extend existing provider implementations or create your own from scratch
|
documentation provider:multicloud
|
It would be extremely helpful to have documentation on extending existing provider implementations or creating new ones from scratch. This would allow developers to customize the infrastructure to meet their specific requirements more easily, and it would also make Booster easier to adopt in specific environments with special compliance requirements.
Currently, there is very little information available on how to do this, as it would require developers to reverse engineering existing implementations, and it can be difficult for developers to figure it out on their own. This lack of documentation has likely been a barrier to adoption for many people.
|
1.0
|
Write documentation on how to extend existing provider implementations or create your own from scratch - It would be extremely helpful to have documentation on extending existing provider implementations or creating new ones from scratch. This would allow developers to customize the infrastructure to meet their specific requirements more easily, and it would also make Booster easier to adopt in specific environments with special compliance requirements.
Currently, there is very little information available on how to do this, as it would require developers to reverse engineering existing implementations, and it can be difficult for developers to figure it out on their own. This lack of documentation has likely been a barrier to adoption for many people.
|
non_process
|
write documentation on how to extend existing provider implementations or create your own from scratch it would be extremely helpful to have documentation on extending existing provider implementations or creating new ones from scratch this would allow developers to customize the infrastructure to meet their specific requirements more easily and it would also make booster easier to adopt in specific environments with special compliance requirements currently there is very little information available on how to do this as it would require developers to reverse engineering existing implementations and it can be difficult for developers to figure it out on their own this lack of documentation has likely been a barrier to adoption for many people
| 0
|
7,271
| 10,425,352,520
|
IssuesEvent
|
2019-09-16 15:15:50
|
SpongePowered/Mixin
|
https://api.github.com/repos/SpongePowered/Mixin
|
closed
|
Add support for SuppressWarnings annotation in Mixin AP to allow silencing nuisance warnings
|
annotation processor enhancement
|
I want to have zero warnings when compiling my code, but currently mixin makes this impossible. This is basically the same issue as #290 but since `@Dynamic` not helping is intentional, it was closed. But this still leaves no solution to silence the warnings.
Consider there is a class `X` (in my case `RenderChunk`) that some coremod (in my case, OptiFine) modifies by adding some fields (`renderChunkNeighboursValid` and `renderChunkNeighbours`), that I need to access within that my mixin.
Adding a `@Shadow` for these fields works perfectly fine when running the code, but mixin AP will show the following error, even with `remap=false`:
```
MixinRenderChunk.java:51: warning: Cannot find target for @Shadow field in net.minecraft.client.renderer.chunk.RenderChunk
@Dynamic @Shadow(remap = false) private RenderChunk[] renderChunkNeighbours;
^
```
While it may not look like a big issue ("it's just warnings"), it makes it very hard to notice the meaningful warnings that I don't want to ignore in the compile output.
|
1.0
|
Add support for SuppressWarnings annotation in Mixin AP to allow silencing nuisance warnings - I want to have zero warnings when compiling my code, but currently mixin makes this impossible. This is basically the same issue as #290 but since `@Dynamic` not helping is intentional, it was closed. But this still leaves no solution to silence the warnings.
Consider there is a class `X` (in my case `RenderChunk`) that some coremod (in my case, OptiFine) modifies by adding some fields (`renderChunkNeighboursValid` and `renderChunkNeighbours`), that I need to access within that my mixin.
Adding a `@Shadow` for these fields works perfectly fine when running the code, but mixin AP will show the following error, even with `remap=false`:
```
MixinRenderChunk.java:51: warning: Cannot find target for @Shadow field in net.minecraft.client.renderer.chunk.RenderChunk
@Dynamic @Shadow(remap = false) private RenderChunk[] renderChunkNeighbours;
^
```
While it may not look like a big issue ("it's just warnings"), it makes it very hard to notice the meaningful warnings that I don't want to ignore in the compile output.
|
process
|
add support for suppresswarnings annotation in mixin ap to allow silencing nuisance warnings i want to have zero warnings when compiling my code but currently mixin makes this impossible this is basically the same issue as but since dynamic not helping is intentional it was closed but this still leaves no solution to silence the warnings consider there is a class x in my case renderchunk that some coremod in my case optifine modifies by adding some fields renderchunkneighboursvalid and renderchunkneighbours that i need to access within that my mixin adding a shadow for these fields works perfectly fine when running the code but mixin ap will show the following error even with remap false mixinrenderchunk java warning cannot find target for shadow field in net minecraft client renderer chunk renderchunk dynamic shadow remap false private renderchunk renderchunkneighbours while it may not look like a big issue it s just warnings it makes it very hard to notice the meaningful warnings that i don t want to ignore in the compile output
| 1
|
767,487
| 26,927,623,433
|
IssuesEvent
|
2023-02-07 14:48:17
|
daisy/ebraille
|
https://api.github.com/repos/daisy/ebraille
|
opened
|
Automatic, converted, and human prepared files will have different levels of quality and expectations
|
use case High Priority content spec metadata
|
I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format.
*Detail*
The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format.
*Proposal*
Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion.
With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
|
1.0
|
Automatic, converted, and human prepared files will have different levels of quality and expectations - I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format.
*Detail*
The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format.
*Proposal*
Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion.
With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
|
non_process
|
automatic converted and human prepared files will have different levels of quality and expectations i am a braille user and my expectations for the quality of file i am receiving will vary based on whether the file was prepared by a transcriber an automatic process or converted from a braille file in an older format detail the best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber next will likely be those converted from an automatic process finally there will be files prepared by converting a braille file in an older format proposal metadata should indicate whether the file was prepared by a human an automatic process or converted from an older braille file if a human it would be good if it indicated the person s level of braille competence professional vs amateur though this aspect needs more thought if an automatic process it would be good if it indicated the file type of the source file as each will have strengths and weaknesses the program used for the automatic process would also be good to indicate finally files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion with this information schools libraries and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes programs and starting files work best for them and their needs
| 0
|
340,408
| 10,272,042,724
|
IssuesEvent
|
2019-08-23 15:27:18
|
KSP-SpaceDock/SpaceDock
|
https://api.github.com/repos/KSP-SpaceDock/SpaceDock
|
opened
|
On the Game page the banner "Browse N more Mods" counts unpublished mods
|
Priority: Low Type: Backend Type: Bug
|
Description says it all
|
1.0
|
On the Game page the banner "Browse N more Mods" counts unpublished mods - Description says it all
|
non_process
|
on the game page the banner browse n more mods counts unpublished mods description says it all
| 0
|
6,944
| 10,112,666,444
|
IssuesEvent
|
2019-07-30 15:07:08
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
PowerShell Steps for ML Anomaly Detection
|
Pri2 assigned-to-author machine-learning/svc product-question team-data-science-process/subsvc triaged
|
Will the 'Invoke-MLAnomalyapi', as shown in the below video, be added to the Az.MachineLearning module in future?
https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443
[](https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2d8a13e5-13f6-ffa2-ba7a-61a04c278293
* Version Independent ID: bbde9854-2484-e8aa-0eb8-8b641706a265
* Content: [Azure Machine Learning Anomaly Detection API - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/apps-anomaly-detection-api)
* Content Source: [articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md](https://github.com/Microsoft/azure-docs/blob/master/articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md)
* Service: **machine-learning**
* Sub-service: **team-data-science-process**
* GitHub Login: @marktab
* Microsoft Alias: **tdsp**
|
1.0
|
PowerShell Steps for ML Anomaly Detection - Will the 'Invoke-MLAnomalyapi', as shown in the below video, be added to the Az.MachineLearning module in future?
https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443
[](https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2d8a13e5-13f6-ffa2-ba7a-61a04c278293
* Version Independent ID: bbde9854-2484-e8aa-0eb8-8b641706a265
* Content: [Azure Machine Learning Anomaly Detection API - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/apps-anomaly-detection-api)
* Content Source: [articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md](https://github.com/Microsoft/azure-docs/blob/master/articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md)
* Service: **machine-learning**
* Sub-service: **team-data-science-process**
* GitHub Login: @marktab
* Microsoft Alias: **tdsp**
|
process
|
powershell steps for ml anomaly detection will the invoke mlanomalyapi as shown in the below video be added to the az machinelearning module in future document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service machine learning sub service team data science process github login marktab microsoft alias tdsp
| 1
|
11,824
| 14,648,348,865
|
IssuesEvent
|
2020-12-27 02:13:47
|
YDongY/Coding
|
https://api.github.com/repos/YDongY/Coding
|
opened
|
02 Process Control | Coding
|
/docs/Shell/ShellScript/02-ProcessControl/ Gitalk
|
https://coding.ydongy.cn/docs/Shell/ShellScript/02-ProcessControl/
流程控制 # 条件判断 # if-then # if 条件/命令; then 指令 fi bash shell 的 if 语句会运行 if 后面的那个命令。如果该命令的退出状态码是 0(该命令成功运行),位于 then 部分的命令就会被执行。如果该命令的退出状态码是其他值,then 部分的命令就不会被执行。fi 语句用来表示 if-then 语句到此结束
#!/bin/bash if pwd ; then echo $(pwd) fi if-then-else # if 条件 ; then 指令1 else 指令2 fi if pwd ; then echo "success." else echo "fail." fi if-elif-then-else # if 条件 then 指令1 elif 条件2 then 指令2 else 指令3 fi #!
|
1.0
|
02 Process Control | Coding - https://coding.ydongy.cn/docs/Shell/ShellScript/02-ProcessControl/
流程控制 # 条件判断 # if-then # if 条件/命令; then 指令 fi bash shell 的 if 语句会运行 if 后面的那个命令。如果该命令的退出状态码是 0(该命令成功运行),位于 then 部分的命令就会被执行。如果该命令的退出状态码是其他值,then 部分的命令就不会被执行。fi 语句用来表示 if-then 语句到此结束
#!/bin/bash if pwd ; then echo $(pwd) fi if-then-else # if 条件 ; then 指令1 else 指令2 fi if pwd ; then echo "success." else echo "fail." fi if-elif-then-else # if 条件 then 指令1 elif 条件2 then 指令2 else 指令3 fi #!
|
process
|
process control coding 流程控制 条件判断 if then if 条件 命令 then 指令 fi bash shell 的 if 语句会运行 if 后面的那个命令。如果该命令的退出状态码是 (该命令成功运行),位于 then 部分的命令就会被执行。如果该命令的退出状态码是其他值,then 部分的命令就不会被执行。fi 语句用来表示 if then 语句到此结束 bin bash if pwd then echo pwd fi if then else if 条件 then else fi if pwd then echo success else echo fail fi if elif then else if 条件 then elif then else fi
| 1
|
678,133
| 23,189,733,062
|
IssuesEvent
|
2022-08-01 11:34:29
|
ita-social-projects/TeachUA
|
https://api.github.com/repos/ita-social-projects/TeachUA
|
opened
|
[News in administration tab] Admin can not change information in any news item without changing the date of the news to the current date or future date
|
bug Backend Priority: Medium
|
**Environment:** Windows 10 Professional, Chrome 96.0.4664.110
**Reproducible:** always
**Build found:** https://speak-ukrainian.org.ua/dev/
**Preconditions**
1. Log in as an administrator on https://speak-ukrainian.org.ua/dev/ (login: [admin@gmail.com](mailto:admin@gmail.com); password: admin)
2. Go to the administration tab - "Новини"
**Steps to reproduce**
1. Click on "Редагувати" button of any news item
3. Change something into current news item and press Save
4. Pay attention to the result of this action
**Actual result**
After pressing Save button an error "date Дата повинна бути сьогоднішньою або майбутньою." occurs
**Expected result**
Admin can change any news item without changing the date of the news

|
1.0
|
[News in administration tab] Admin can not change information in any news item without changing the date of the news to the current date or future date - **Environment:** Windows 10 Professional, Chrome 96.0.4664.110
**Reproducible:** always
**Build found:** https://speak-ukrainian.org.ua/dev/
**Preconditions**
1. Log in as an administrator on https://speak-ukrainian.org.ua/dev/ (login: [admin@gmail.com](mailto:admin@gmail.com); password: admin)
2. Go to the administration tab - "Новини"
**Steps to reproduce**
1. Click on "Редагувати" button of any news item
3. Change something into current news item and press Save
4. Pay attention to the result of this action
**Actual result**
After pressing Save button an error "date Дата повинна бути сьогоднішньою або майбутньою." occurs
**Expected result**
Admin can change any news item without changing the date of the news

|
non_process
|
admin can not change information in any news item without changing the date of the news to the current date or future date environment windows professional chrome reproducible always build found preconditions log in as an administrator on login mailto admin gmail com password admin go to the administration tab новини steps to reproduce click on редагувати button of any news item change something into current news item and press save pay attention to the result of this action actual result after pressing save button an error date дата повинна бути сьогоднішньою або майбутньою occurs expected result admin can change any news item without changing the date of the news
| 0
|
5,435
| 8,299,549,511
|
IssuesEvent
|
2018-09-21 03:35:31
|
flutterchina/dio
|
https://api.github.com/repos/flutterchina/dio
|
closed
|
Content size exceeds specified contentLength
|
processing
|
## Steps to Reproduce
- 请求方式
```java
Map<String, dynamic> data
dio.post(path, data: data);
```
- 请求失败
```java
requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: 测试测试测试, courseId: 110, score: 3, anonymous: 1, evaluateObject: 2, timestamp: 1537426578634}
```
如果我把content换成英文,则请求成功。
- 请求成功
```java
requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: test, courseId: 110, score: 5, anonymous: 1, evaluateObject: 1, timestamp: 1537426565722}
```
## Logs
```
DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"æµè¯æµè¯æµè¯","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}]
I/flutter (14570): DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"æµè¯æµè¯æµè¯","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}]
```
|
1.0
|
Content size exceeds specified contentLength - ## Steps to Reproduce
- 请求方式
```java
Map<String, dynamic> data
dio.post(path, data: data);
```
- 请求失败
```java
requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: 测试测试测试, courseId: 110, score: 3, anonymous: 1, evaluateObject: 2, timestamp: 1537426578634}
```
如果我把content换成英文,则请求成功。
- 请求成功
```java
requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: test, courseId: 110, score: 5, anonymous: 1, evaluateObject: 1, timestamp: 1537426565722}
```
## Logs
```
DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"æµè¯æµè¯æµè¯","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}]
I/flutter (14570): DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"æµè¯æµè¯æµè¯","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}]
```
|
process
|
content size exceeds specified contentlength steps to reproduce 请求方式 java map data dio post path data data 请求失败 java requestpost host path backend student stuapp evaluate data content 测试测试测试 courseid score anonymous evaluateobject timestamp 如果我把content换成英文,则请求成功。 请求成功 java requestpost host path backend student stuapp evaluate data content test courseid score anonymous evaluateobject timestamp logs dioerror httpexception content size exceeds specified contentlength bytes written while expected i flutter dioerror httpexception content size exceeds specified contentlength bytes written while expected
| 1
|
9,992
| 13,039,523,833
|
IssuesEvent
|
2020-07-28 16:54:01
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
[tests] Tests in tables/automl/dataset_test.py expect to have access to gs://python-docs-samples-tests-automl-tables-test
|
api: automl priority: p2 type: process
|
## In which file did you encounter the issue?
tables/automl/dataset_test.py
## Describe the issue
All tests are failing with error:
AccessDeniedException: 403 xxx@xxx.com does not have storage.objects.list access to the Google Cloud Storage bucket.
It's not clear what account should I use for tests. All that said in AUTHORING_GUIDE.MD is that I need to setup my own project. But this GCS absolute bucket path is hard-coded inside tests file. How should I get access to this bucket?
|
1.0
|
[tests] Tests in tables/automl/dataset_test.py expect to have access to gs://python-docs-samples-tests-automl-tables-test - ## In which file did you encounter the issue?
tables/automl/dataset_test.py
## Describe the issue
All tests are failing with error:
AccessDeniedException: 403 xxx@xxx.com does not have storage.objects.list access to the Google Cloud Storage bucket.
It's not clear what account should I use for tests. All that said in AUTHORING_GUIDE.MD is that I need to setup my own project. But this GCS absolute bucket path is hard-coded inside tests file. How should I get access to this bucket?
|
process
|
tests in tables automl dataset test py expect to have access to gs python docs samples tests automl tables test in which file did you encounter the issue tables automl dataset test py describe the issue all tests are failing with error accessdeniedexception xxx xxx com does not have storage objects list access to the google cloud storage bucket it s not clear what account should i use for tests all that said in authoring guide md is that i need to setup my own project but this gcs absolute bucket path is hard coded inside tests file how should i get access to this bucket
| 1
|
205,136
| 23,299,717,965
|
IssuesEvent
|
2022-08-07 05:57:10
|
TheFloAnd/eventlisting-laravel
|
https://api.github.com/repos/TheFloAnd/eventlisting-laravel
|
closed
|
laravel-mix-6.0.43.tgz: 2 vulnerabilities (highest severity is: 7.5) - autoclosed
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel-mix-6.0.43.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/terser/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-25858](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | detected in multiple dependencies | Transitive | N/A | ❌ |
| [CVE-2021-35065](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-5.1.2.tgz | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25858</summary>
### Vulnerable Libraries - <b>terser-4.8.0.tgz</b>, <b>terser-5.12.1.tgz</b></p>
<p>
### <b>terser-4.8.0.tgz</b></p>
<p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p>
<p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-4.8.0.tgz">https://registry.npmjs.org/terser/-/terser-4.8.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/html-minifier-terser/node_modules/terser/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-6.0.43.tgz (Root Library)
- html-loader-1.3.2.tgz
- html-minifier-terser-5.1.1.tgz
- :x: **terser-4.8.0.tgz** (Vulnerable Library)
### <b>terser-5.12.1.tgz</b></p>
<p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p>
<p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-5.12.1.tgz">https://registry.npmjs.org/terser/-/terser-5.12.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/terser/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-6.0.43.tgz (Root Library)
- :x: **terser-5.12.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package terser before 4.8.1, from 5.0.0 and before 5.14.2 are vulnerable to Regular Expression Denial of Service (ReDoS) due to insecure usage of regular expressions.
<p>Publish Date: 2022-07-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858>CVE-2022-25858</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858</a></p>
<p>Release Date: 2022-07-15</p>
<p>Fix Resolution: terser - 4.8.1,5.14.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-35065</summary>
### Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-6.0.43.tgz (Root Library)
- chokidar-3.5.3.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
laravel-mix-6.0.43.tgz: 2 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel-mix-6.0.43.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/terser/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-25858](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | detected in multiple dependencies | Transitive | N/A | ❌ |
| [CVE-2021-35065](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-5.1.2.tgz | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25858</summary>
### Vulnerable Libraries - <b>terser-4.8.0.tgz</b>, <b>terser-5.12.1.tgz</b></p>
<p>
### <b>terser-4.8.0.tgz</b></p>
<p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p>
<p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-4.8.0.tgz">https://registry.npmjs.org/terser/-/terser-4.8.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/html-minifier-terser/node_modules/terser/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-6.0.43.tgz (Root Library)
- html-loader-1.3.2.tgz
- html-minifier-terser-5.1.1.tgz
- :x: **terser-4.8.0.tgz** (Vulnerable Library)
### <b>terser-5.12.1.tgz</b></p>
<p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p>
<p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-5.12.1.tgz">https://registry.npmjs.org/terser/-/terser-5.12.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/terser/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-6.0.43.tgz (Root Library)
- :x: **terser-5.12.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package terser before 4.8.1, from 5.0.0 and before 5.14.2 are vulnerable to Regular Expression Denial of Service (ReDoS) due to insecure usage of regular expressions.
<p>Publish Date: 2022-07-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858>CVE-2022-25858</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858</a></p>
<p>Release Date: 2022-07-15</p>
<p>Fix Resolution: terser - 4.8.1,5.14.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-35065</summary>
### Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-6.0.43.tgz (Root Library)
- chokidar-3.5.3.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
laravel mix tgz vulnerabilities highest severity is autoclosed vulnerable library laravel mix tgz path to dependency file package json path to vulnerable library node modules terser package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high detected in multiple dependencies transitive n a high glob parent tgz transitive n a details cve vulnerable libraries terser tgz terser tgz terser tgz javascript parser mangler compressor and beautifier toolkit for library home page a href path to dependency file package json path to vulnerable library node modules html minifier terser node modules terser package json dependency hierarchy laravel mix tgz root library html loader tgz html minifier terser tgz x terser tgz vulnerable library terser tgz javascript parser mangler compressor and beautifier toolkit for library home page a href path to dependency file package json path to vulnerable library node modules terser package json dependency hierarchy laravel mix tgz root library x terser tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package terser before from and before are vulnerable to regular expression denial of service redos due to insecure usage of regular expressions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution terser step up your open source security game with mend cve vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy laravel mix tgz root library chokidar tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with mend
| 0
|
279,751
| 24,252,594,201
|
IssuesEvent
|
2022-09-27 15:12:20
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Test badge on TreeView
|
testplan-item
|
Refs: https://github.com/microsoft/vscode/issues/62783
- [x] anyOS @meganrogge
- [x] anyOS @bpasero
Complexity: 4
Roles: Developer
[Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23161780%0A%0A&assignees=alexr00)
---
We have newly finalized API for badges on TreeViews: https://github.com/microsoft/vscode/blob/9caff0557b9ed17a6ccb44b2f18c1f6915f2eb85/src/vscode-dts/vscode.d.ts#L10484-L10488
This badge shows as a circle with a number on the view's view container.
To test:
- Read the inline documentation for the API and verify that it's clear.
- Try out the API in a tree view. For example, in the tree-view-sample: https://github.com/microsoft/vscode-extension-samples/blob/31d3d4920b3fae52e03a185e434f51ee81965515/tree-view-sample/src/testView.ts#L6-L7
- Verify that you can set the badge number value.
- Verify that you can update the badge number value.
- Verify that you can cause the badge to hide by unsetting the badge number value.
|
1.0
|
Test badge on TreeView - Refs: https://github.com/microsoft/vscode/issues/62783
- [x] anyOS @meganrogge
- [x] anyOS @bpasero
Complexity: 4
Roles: Developer
[Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23161780%0A%0A&assignees=alexr00)
---
We have newly finalized API for badges on TreeViews: https://github.com/microsoft/vscode/blob/9caff0557b9ed17a6ccb44b2f18c1f6915f2eb85/src/vscode-dts/vscode.d.ts#L10484-L10488
This badge shows as a circle with a number on the view's view container.
To test:
- Read the inline documentation for the API and verify that it's clear.
- Try out the API in a tree view. For example, in the tree-view-sample: https://github.com/microsoft/vscode-extension-samples/blob/31d3d4920b3fae52e03a185e434f51ee81965515/tree-view-sample/src/testView.ts#L6-L7
- Verify that you can set the badge number value.
- Verify that you can update the badge number value.
- Verify that you can cause the badge to hide by unsetting the badge number value.
|
non_process
|
test badge on treeview refs anyos meganrogge anyos bpasero complexity roles developer we have newly finalized api for badges on treeviews this badge shows as a circle with a number on the view s view container to test read the inline documentation for the api and verify that it s clear try out the api in a tree view for example in the tree view sample verify that you can set the badge number value verify that you can update the badge number value verify that you can cause the badge to hide by unsetting the badge number value
| 0
|
16,093
| 20,262,151,763
|
IssuesEvent
|
2022-02-15 08:39:47
|
sillsdev/silnlp
|
https://api.github.com/repos/sillsdev/silnlp
|
opened
|
The source vocab for a multilingual parent model is not correctly transferred to the child model
|
bug pipeline 3: preprocess
|
The source vocab for a multilingual model will contain special target language tags. When a child model is finetuned from a multilingual model, it will need to add a new target language tag to the vocab. Because of this, the preprocess script will build a new sentencepiece model, because the child vocab contains tokens that do not exist in the parent vocab. In this case, we should preserve the parent vocab while still adding the new target language tag.
|
1.0
|
The source vocab for a multilingual parent model is not correctly transferred to the child model - The source vocab for a multilingual model will contain special target language tags. When a child model is finetuned from a multilingual model, it will need to add a new target language tag to the vocab. Because of this, the preprocess script will build a new sentencepiece model, because the child vocab contains tokens that do not exist in the parent vocab. In this case, we should preserve the parent vocab while still adding the new target language tag.
|
process
|
the source vocab for a multilingual parent model is not correctly transferred to the child model the source vocab for a multilingual model will contain special target language tags when a child model is finetuned from a multilingual model it will need to add a new target language tag to the vocab because of this the preprocess script will build a new sentencepiece model because the child vocab contains tokens that do not exist in the parent vocab in this case we should preserve the parent vocab while still adding the new target language tag
| 1
|
160,794
| 13,797,237,431
|
IssuesEvent
|
2020-10-09 21:22:51
|
rrousselGit/river_pod
|
https://api.github.com/repos/rrousselGit/river_pod
|
closed
|
Usage of watch. Documentation shows 2 different usage of watch
|
documentation
|
Documentation shows 2 different usage of watch:
`watch(xxxProvider.state)` and `watch(xxxProvider).state` , is there any difference? clarify when to use which?
|
1.0
|
Usage of watch. Documentation shows 2 different usage of watch - Documentation shows 2 different usage of watch:
`watch(xxxProvider.state)` and `watch(xxxProvider).state` , is there any difference? clarify when to use which?
|
non_process
|
usage of watch documentation shows different usage of watch documentation shows different usage of watch watch xxxprovider state and watch xxxprovider state is there any difference clarify when to use which
| 0
|
7,739
| 10,862,521,073
|
IssuesEvent
|
2019-11-14 13:28:38
|
bisq-network/bisq
|
https://api.github.com/repos/bisq-network/bisq
|
closed
|
Buying/selling grin not possible through bisq?
|
in:altcoins in:trade-process was:dropped
|
Grin was added to bisq in https://github.com/bisq-network/bisq/pull/2217 but as far as I can tell it's not possible to (reliably, or maybe at all) buy/sell grin through bisq for two reasons:
1. grin cannot be transferred without off-chain interaction between accounts in question (see e.g. https://blog.blockcypher.com/mining-grin-on-grinmint-how-to-request-payouts-1593835181eb). Easiest way with bisq seems to be that buyer of grin uses ip address as their grin address in bisq so that seller can send grin by connecting grin wallet/node directly to buyer's wallet/node. This feels awkward and might also be security issue since everything else in bisq runs over tor.
2. currently it's not possible to prove that grin transfer has taken place: https://github.com/mimblewimble/grin/issues/2336
Was grin added prematurely to bisq?
|
1.0
|
Buying/selling grin not possible through bisq? - Grin was added to bisq in https://github.com/bisq-network/bisq/pull/2217 but as far as I can tell it's not possible to (reliably, or maybe at all) buy/sell grin through bisq for two reasons:
1. grin cannot be transferred without off-chain interaction between accounts in question (see e.g. https://blog.blockcypher.com/mining-grin-on-grinmint-how-to-request-payouts-1593835181eb). Easiest way with bisq seems to be that buyer of grin uses ip address as their grin address in bisq so that seller can send grin by connecting grin wallet/node directly to buyer's wallet/node. This feels awkward and might also be security issue since everything else in bisq runs over tor.
2. currently it's not possible to prove that grin transfer has taken place: https://github.com/mimblewimble/grin/issues/2336
Was grin added prematurely to bisq?
|
process
|
buying selling grin not possible through bisq grin was added to bisq in but as far as i can tell it s not possible to reliably or maybe at all buy sell grin through bisq for two reasons grin cannot be transferred without off chain interaction between accounts in question see e g easiest way with bisq seems to be that buyer of grin uses ip address as their grin address in bisq so that seller can send grin by connecting grin wallet node directly to buyer s wallet node this feels awkward and might also be security issue since everything else in bisq runs over tor currently it s not possible to prove that grin transfer has taken place was grin added prematurely to bisq
| 1
|
291,606
| 21,931,781,228
|
IssuesEvent
|
2022-05-23 10:22:00
|
typeorm/typeorm
|
https://api.github.com/repos/typeorm/typeorm
|
closed
|
connection-api falsely imports getEntityManager instead of getManager
|
documentation requires triage
|
<!--
Please follow the template. If you don't, your issue may be closed.
Have a question? This is the TypeORM issue tracker - and not the right place
for general support or questions. Instead, check the "Support" Documentation
on the best places to ask questions!
https://github.com/typeorm/typeorm/blob/master/docs/support.md
-->
## Documentation Issue
### What was unclear or otherwise insufficient?
In connection-api, there's a usage of getEntityManager, but it should be changed to getManager
<!--
If relevant, Please be clear about the documentation file,
as well as the location within the file. Link to the documentation
in the repository.
If the page does not exist, please be clear why a new documentation
section is needed.
-->
### Recommended Fix
PR#8504
<!--
How should we fix this documentation issue?
Should we add examples, clarify the language, or drop the page entirely?
-->
### Additional Context
<!--
Add any other context about the documentation issue here.
-->
### Are you willing to resolve this issue by submitting a Pull Request?
<!-- Put "✅" (:white_check_mark:) to one of these options, left "✖️" (:heavy_multiplication_x:) others: -->
- ✅ Yes, I have the time, and I know how to start.
- ✖️ Yes, I have the time, but I don't know how to start. I would need guidance.
- ✖️ No, I don’t have the time, but I can support (using donations) development.
- ✖️ No, I don’t have the time and I’m okay to wait for the community / maintainers to resolve this issue.
<!--
Remember that first-time contributors are welcome! 🙌
👋 Have a great day and thank you for the documentation problem report!
-->
|
1.0
|
connection-api falsely imports getEntityManager instead of getManager - <!--
Please follow the template. If you don't, your issue may be closed.
Have a question? This is the TypeORM issue tracker - and not the right place
for general support or questions. Instead, check the "Support" Documentation
on the best places to ask questions!
https://github.com/typeorm/typeorm/blob/master/docs/support.md
-->
## Documentation Issue
### What was unclear or otherwise insufficient?
In connection-api, there's a usage of getEntityManager, but it should be changed to getManager
<!--
If relevant, Please be clear about the documentation file,
as well as the location within the file. Link to the documentation
in the repository.
If the page does not exist, please be clear why a new documentation
section is needed.
-->
### Recommended Fix
PR#8504
<!--
How should we fix this documentation issue?
Should we add examples, clarify the language, or drop the page entirely?
-->
### Additional Context
<!--
Add any other context about the documentation issue here.
-->
### Are you willing to resolve this issue by submitting a Pull Request?
<!-- Put "✅" (:white_check_mark:) to one of these options, left "✖️" (:heavy_multiplication_x:) others: -->
- ✅ Yes, I have the time, and I know how to start.
- ✖️ Yes, I have the time, but I don't know how to start. I would need guidance.
- ✖️ No, I don’t have the time, but I can support (using donations) development.
- ✖️ No, I don’t have the time and I’m okay to wait for the community / maintainers to resolve this issue.
<!--
Remember that first-time contributors are welcome! 🙌
👋 Have a great day and thank you for the documentation problem report!
-->
|
non_process
|
connection api falsely imports getentitymanager instead of getmanager please follow the template if you don t your issue may be closed have a question this is the typeorm issue tracker and not the right place for general support or questions instead check the support documentation on the best places to ask questions documentation issue what was unclear or otherwise insufficient in connection api there s a usage of getentitymanager but it should be changed to getmanager if relevant please be clear about the documentation file as well as the location within the file link to the documentation in the repository if the page does not exist please be clear why a new documentation section is needed recommended fix pr how should we fix this documentation issue should we add examples clarify the language or drop the page entirely additional context add any other context about the documentation issue here are you willing to resolve this issue by submitting a pull request ✅ yes i have the time and i know how to start ✖️ yes i have the time but i don t know how to start i would need guidance ✖️ no i don’t have the time but i can support using donations development ✖️ no i don’t have the time and i’m okay to wait for the community maintainers to resolve this issue remember that first time contributors are welcome 🙌 👋 have a great day and thank you for the documentation problem report
| 0
|
90,661
| 26,162,176,809
|
IssuesEvent
|
2022-12-31 18:33:58
|
sandboxie-plus/Sandboxie
|
https://api.github.com/repos/sandboxie-plus/Sandboxie
|
closed
|
SandMan: Explorer context menu does not work directly after a clean installation on Windows 11
|
fixed in next build Issue reproduced Win 11 SbieDll
|
### Describe what you noticed and did
With a clean reinstallation, the **Explorer context menu** (shell integration) usually only works after a restart of Windows.
### Clean VM (running **Windows 11**, 21H2, 64-bit):
1.) Download and install the latest version of Sandboxie-Plus (v1.3.4, 64-bit).
2.) Click through the Setup Wizard "Personally", leaving everything at default values.
3.) Wait until the entry "Open Sandboxed" is added to the Explorer context menu.
4.) Run e.g. "Microsoft Edge" in the DefaultBox via the Explorer context menu.
5.) The execution does not work, but the following error messages are displayed:

`Start.exe (1504): SBIE2112 Object is not accessible: \Sessions\1\BaseNamedObjects\SM0:1504:120:WilError_03, call CreateMutant (C0000022) access=001F0001 initialized=1`
After restarting Windows, it then works as it should.
~~**Note: Update 1 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1259720113**~~
~~**Note: Update 2 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1282416049**~~
**Note: Update 3 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1328778347**
### How often did you encounter it so far?
With each clean reinstall of Sandboxie Plus.
### Affected program
Not relevant
### Download link
Not relevant
### Where is the program located?
Not relevant to my request.
### Expected behavior
The Explorer context menu should work correctly immediately (as under SbieCtrl) or the user should be prompted to restart.
### What is your Windows edition and version?
Windows 11 Pro, 21H2, 64-bit
### In which Windows account you have this problem?
User account with UAC protection set to Always notify.
### Please mention any installed security software
Microsoft Defender Antivirus
### What version of Sandboxie are you running?
1.3.2-1.6.3
### Is it a new installation of Sandboxie?
I recently did a new clean installation.
### Is it a regression?
_No response_
### In which sandbox type you have this problem?
Not relevant to my request.
### Can you reproduce this problem on an empty sandbox?
Not relevant to my request.
### Did you previously enable some security policy settings outside Sandboxie?
_No response_
### Crash dump
_No response_
### Trace log
_No response_
### Sandboxie.ini configuration
_No response_
|
1.0
|
SandMan: Explorer context menu does not work directly after a clean installation on Windows 11 - ### Describe what you noticed and did
With a clean reinstallation, the **Explorer context menu** (shell integration) usually only works after a restart of Windows.
### Clean VM (running **Windows 11**, 21H2, 64-bit):
1.) Download and install the latest version of Sandboxie-Plus (v1.3.4, 64-bit).
2.) Click through the Setup Wizard "Personally", leaving everything at default values.
3.) Wait until the entry "Open Sandboxed" is added to the Explorer context menu.
4.) Run e.g. "Microsoft Edge" in the DefaultBox via the Explorer context menu.
5.) The execution does not work, but the following error messages are displayed:

`Start.exe (1504): SBIE2112 Object is not accessible: \Sessions\1\BaseNamedObjects\SM0:1504:120:WilError_03, call CreateMutant (C0000022) access=001F0001 initialized=1`
After restarting Windows, it then works as it should.
~~**Note: Update 1 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1259720113**~~
~~**Note: Update 2 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1282416049**~~
**Note: Update 3 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1328778347**
### How often did you encounter it so far?
With each clean reinstall of Sandboxie Plus.
### Affected program
Not relevant
### Download link
Not relevant
### Where is the program located?
Not relevant to my request.
### Expected behavior
The Explorer context menu should work correctly immediately (as under SbieCtrl) or the user should be prompted to restart.
### What is your Windows edition and version?
Windows 11 Pro, 21H2, 64-bit
### In which Windows account you have this problem?
User account with UAC protection set to Always notify.
### Please mention any installed security software
Microsoft Defender Antivirus
### What version of Sandboxie are you running?
1.3.2-1.6.3
### Is it a new installation of Sandboxie?
I recently did a new clean installation.
### Is it a regression?
_No response_
### In which sandbox type you have this problem?
Not relevant to my request.
### Can you reproduce this problem on an empty sandbox?
Not relevant to my request.
### Did you previously enable some security policy settings outside Sandboxie?
_No response_
### Crash dump
_No response_
### Trace log
_No response_
### Sandboxie.ini configuration
_No response_
|
non_process
|
sandman explorer context menu does not work directly after a clean installation on windows describe what you noticed and did with a clean reinstallation the explorer context menu shell integration usually only works after a restart of windows clean vm running windows bit download and install the latest version of sandboxie plus bit click through the setup wizard personally leaving everything at default values wait until the entry open sandboxed is added to the explorer context menu run e g microsoft edge in the defaultbox via the explorer context menu the execution does not work but the following error messages are displayed start exe object is not accessible sessions basenamedobjects wilerror call createmutant access initialized after restarting windows it then works as it should note update note update note update how often did you encounter it so far with each clean reinstall of sandboxie plus affected program not relevant download link not relevant where is the program located not relevant to my request expected behavior the explorer context menu should work correctly immediately as under sbiectrl or the user should be prompted to restart what is your windows edition and version windows pro bit in which windows account you have this problem user account with uac protection set to always notify please mention any installed security software microsoft defender antivirus what version of sandboxie are you running is it a new installation of sandboxie i recently did a new clean installation is it a regression no response in which sandbox type you have this problem not relevant to my request can you reproduce this problem on an empty sandbox not relevant to my request did you previously enable some security policy settings outside sandboxie no response crash dump no response trace log no response sandboxie ini configuration no response
| 0
|
15,158
| 18,909,998,106
|
IssuesEvent
|
2021-11-16 13:12:51
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Better error message if using `TEXT` or `BLOB` in MySQL @id/@index/@unique
|
process/candidate topic: indexes team/migrations topic: extendedIndexes
|
We now say this is not allowed. We should point to the correct preview feature and mention about using the `length` argument for indexes to work with TEXT or BLOB fields.
|
1.0
|
Better error message if using `TEXT` or `BLOB` in MySQL @id/@index/@unique - We now say this is not allowed. We should point to the correct preview feature and mention about using the `length` argument for indexes to work with TEXT or BLOB fields.
|
process
|
better error message if using text or blob in mysql id index unique we now say this is not allowed we should point to the correct preview feature and mention about using the length argument for indexes to work with text or blob fields
| 1
|
52,493
| 27,592,009,460
|
IssuesEvent
|
2023-03-09 01:33:07
|
keras-team/keras
|
https://api.github.com/repos/keras-team/keras
|
closed
|
Can't excute ConvNeXt with input_tensor
|
type:bug/performance stat:awaiting response from contributor stale
|
When keras.applications.ConvNeXt is executed with input_tensor, it is not executed.
The cause seems to be a problem that occurs as it changes to the form of '[input_tensor]' when passing utils.layer_utils.get_source_inputs.
(When using input_shape, there is no problem.)
This problem was identified in versions tensorflow==2.10 and keras==2.10.
Check please.
the following error message:
```
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
Input In [79], in <cell line: 2>()
1 x = tf.keras.layers.Input(shape = (512,512,3))
----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = True, include_top = False, weights = None)
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
562 @keras_export(
563 "keras.applications.convnext.ConvNeXtTiny",
564 "keras.applications.ConvNeXtTiny",
(...)
575 classifier_activation="softmax",
576 ):
--> 577 return ConvNeXt(
578 depths=MODEL_CONFIGS["tiny"]["depths"],
579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"],
580 drop_path_rate=0.0,
581 layer_scale_init_value=1e-6,
582 default_size=MODEL_CONFIGS["tiny"]["default_size"],
583 model_name=model_name,
584 include_top=include_top,
585 include_preprocessing=include_preprocessing,
586 weights=weights,
587 input_tensor=input_tensor,
588 input_shape=input_shape,
589 pooling=pooling,
590 classes=classes,
591 classifier_activation=classifier_activation,
592 )
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:461, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
459 num_channels = input_shape[channel_axis - 1]
460 if num_channels == 3:
--> 461 x = PreStem(name=model_name)(x)
463 # Stem block.
464 stem = sequential.Sequential(
465 [
466 layers.Conv2D(
(...)
476 name=model_name + "_stem",
477 )
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:313, in PreStem.<locals>.apply(x)
312 def apply(x):
--> 313 x = layers.Normalization(
314 mean=[0.485 * 255, 0.456 * 255, 0.406 * 255],
315 variance=[
316 (0.229 * 255) ** 2,
317 (0.224 * 255) ** 2,
318 (0.225 * 255) ** 2,
319 ],
320 name=name + "_prestem_normalization",
321 )(x)
322 return x
File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs)
67 filtered_tb = _process_traceback_frames(e.__traceback__)
68 # To get the full stack trace, call:
69 # `tf.debugging.disable_traceback_filtering()`
---> 70 raise e.with_traceback(filtered_tb) from None
71 finally:
72 del filtered_tb
File /opt/anaconda3/lib/python3.9/site-packages/keras/layers/preprocessing/normalization.py:166, in Normalization.build(self, input_shape)
161 super().build(input_shape)
163 if isinstance(input_shape, (list, tuple)) and all(
164 isinstance(shape, tf.TensorShape) for shape in input_shape
165 ):
--> 166 raise ValueError(
167 "Normalization only accepts a single input. If you are "
168 "passing a python list or tuple as a single input, "
169 "please convert to a numpy array or `tf.Tensor`."
170 )
172 input_shape = tf.TensorShape(input_shape).as_list()
173 ndim = len(input_shape)
ValueError: Normalization only accepts a single input. If you are passing a python list or tuple as a single input, please convert to a numpy array or `tf.Tensor`.
```
the following error message without include_preprocessing:
```
WARNING:tensorflow:Layers in a Sequential model should only have a single input tensor. Received: inputs=[<tf.Tensor 'Placeholder:0' shape=(None, 512, 512, 3) dtype=float32>]. Consider rewriting this model with the Functional API.
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
Input In [81], in <cell line: 2>()
1 x = tf.keras.layers.Input(shape = (512,512,3))
----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = False, include_top = False, weights = None)
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
562 @keras_export(
563 "keras.applications.convnext.ConvNeXtTiny",
564 "keras.applications.ConvNeXtTiny",
(...)
575 classifier_activation="softmax",
576 ):
--> 577 return ConvNeXt(
578 depths=MODEL_CONFIGS["tiny"]["depths"],
579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"],
580 drop_path_rate=0.0,
581 layer_scale_init_value=1e-6,
582 default_size=MODEL_CONFIGS["tiny"]["default_size"],
583 model_name=model_name,
584 include_top=include_top,
585 include_preprocessing=include_preprocessing,
586 weights=weights,
587 input_tensor=input_tensor,
588 input_shape=input_shape,
589 pooling=pooling,
590 classes=classes,
591 classifier_activation=classifier_activation,
592 )
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:514, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
512 num_convnext_blocks = 4
513 for i in range(num_convnext_blocks):
--> 514 x = downsample_layers[i](x)
515 for j in range(depths[i]):
516 x = ConvNeXtBlock(
517 projection_dim=projection_dims[i],
518 drop_path_rate=depth_drop_rates[cur + j],
519 layer_scale_init_value=layer_scale_init_value,
520 name=model_name + f"_stage_{i}_block_{j}",
521 )(x)
File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs)
67 filtered_tb = _process_traceback_frames(e.__traceback__)
68 # To get the full stack trace, call:
69 # `tf.debugging.disable_traceback_filtering()`
---> 70 raise e.with_traceback(filtered_tb) from None
71 finally:
72 del filtered_tb
File <string>:3, in raise_from(value, from_value)
TypeError: Exception encountered when calling layer "convnext_tiny_stem" (type Sequential).
Dimension value must be integer or None or have an __index__ method, got value 'TensorShape([None, 512, 512, 3])' with type '<class 'tensorflow.python.framework.tensor_shape.TensorShape'>'
Call arguments received by layer "convnext_tiny_stem" (type Sequential):
• inputs=['tf.Tensor(shape=(None, 512, 512, 3), dtype=float32)']
• training=False
• mask=None
```
|
True
|
Can't excute ConvNeXt with input_tensor - When keras.applications.ConvNeXt is executed with input_tensor, it is not executed.
The cause seems to be a problem that occurs as it changes to the form of '[input_tensor]' when passing utils.layer_utils.get_source_inputs.
(When using input_shape, there is no problem.)
This problem was identified in versions tensorflow==2.10 and keras==2.10.
Check please.
the following error message:
```
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
Input In [79], in <cell line: 2>()
1 x = tf.keras.layers.Input(shape = (512,512,3))
----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = True, include_top = False, weights = None)
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
562 @keras_export(
563 "keras.applications.convnext.ConvNeXtTiny",
564 "keras.applications.ConvNeXtTiny",
(...)
575 classifier_activation="softmax",
576 ):
--> 577 return ConvNeXt(
578 depths=MODEL_CONFIGS["tiny"]["depths"],
579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"],
580 drop_path_rate=0.0,
581 layer_scale_init_value=1e-6,
582 default_size=MODEL_CONFIGS["tiny"]["default_size"],
583 model_name=model_name,
584 include_top=include_top,
585 include_preprocessing=include_preprocessing,
586 weights=weights,
587 input_tensor=input_tensor,
588 input_shape=input_shape,
589 pooling=pooling,
590 classes=classes,
591 classifier_activation=classifier_activation,
592 )
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:461, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
459 num_channels = input_shape[channel_axis - 1]
460 if num_channels == 3:
--> 461 x = PreStem(name=model_name)(x)
463 # Stem block.
464 stem = sequential.Sequential(
465 [
466 layers.Conv2D(
(...)
476 name=model_name + "_stem",
477 )
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:313, in PreStem.<locals>.apply(x)
312 def apply(x):
--> 313 x = layers.Normalization(
314 mean=[0.485 * 255, 0.456 * 255, 0.406 * 255],
315 variance=[
316 (0.229 * 255) ** 2,
317 (0.224 * 255) ** 2,
318 (0.225 * 255) ** 2,
319 ],
320 name=name + "_prestem_normalization",
321 )(x)
322 return x
File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs)
67 filtered_tb = _process_traceback_frames(e.__traceback__)
68 # To get the full stack trace, call:
69 # `tf.debugging.disable_traceback_filtering()`
---> 70 raise e.with_traceback(filtered_tb) from None
71 finally:
72 del filtered_tb
File /opt/anaconda3/lib/python3.9/site-packages/keras/layers/preprocessing/normalization.py:166, in Normalization.build(self, input_shape)
161 super().build(input_shape)
163 if isinstance(input_shape, (list, tuple)) and all(
164 isinstance(shape, tf.TensorShape) for shape in input_shape
165 ):
--> 166 raise ValueError(
167 "Normalization only accepts a single input. If you are "
168 "passing a python list or tuple as a single input, "
169 "please convert to a numpy array or `tf.Tensor`."
170 )
172 input_shape = tf.TensorShape(input_shape).as_list()
173 ndim = len(input_shape)
ValueError: Normalization only accepts a single input. If you are passing a python list or tuple as a single input, please convert to a numpy array or `tf.Tensor`.
```
the following error message without include_preprocessing:
```
WARNING:tensorflow:Layers in a Sequential model should only have a single input tensor. Received: inputs=[<tf.Tensor 'Placeholder:0' shape=(None, 512, 512, 3) dtype=float32>]. Consider rewriting this model with the Functional API.
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
Input In [81], in <cell line: 2>()
1 x = tf.keras.layers.Input(shape = (512,512,3))
----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = False, include_top = False, weights = None)
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
562 @keras_export(
563 "keras.applications.convnext.ConvNeXtTiny",
564 "keras.applications.ConvNeXtTiny",
(...)
575 classifier_activation="softmax",
576 ):
--> 577 return ConvNeXt(
578 depths=MODEL_CONFIGS["tiny"]["depths"],
579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"],
580 drop_path_rate=0.0,
581 layer_scale_init_value=1e-6,
582 default_size=MODEL_CONFIGS["tiny"]["default_size"],
583 model_name=model_name,
584 include_top=include_top,
585 include_preprocessing=include_preprocessing,
586 weights=weights,
587 input_tensor=input_tensor,
588 input_shape=input_shape,
589 pooling=pooling,
590 classes=classes,
591 classifier_activation=classifier_activation,
592 )
File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:514, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation)
512 num_convnext_blocks = 4
513 for i in range(num_convnext_blocks):
--> 514 x = downsample_layers[i](x)
515 for j in range(depths[i]):
516 x = ConvNeXtBlock(
517 projection_dim=projection_dims[i],
518 drop_path_rate=depth_drop_rates[cur + j],
519 layer_scale_init_value=layer_scale_init_value,
520 name=model_name + f"_stage_{i}_block_{j}",
521 )(x)
File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs)
67 filtered_tb = _process_traceback_frames(e.__traceback__)
68 # To get the full stack trace, call:
69 # `tf.debugging.disable_traceback_filtering()`
---> 70 raise e.with_traceback(filtered_tb) from None
71 finally:
72 del filtered_tb
File <string>:3, in raise_from(value, from_value)
TypeError: Exception encountered when calling layer "convnext_tiny_stem" (type Sequential).
Dimension value must be integer or None or have an __index__ method, got value 'TensorShape([None, 512, 512, 3])' with type '<class 'tensorflow.python.framework.tensor_shape.TensorShape'>'
Call arguments received by layer "convnext_tiny_stem" (type Sequential):
• inputs=['tf.Tensor(shape=(None, 512, 512, 3), dtype=float32)']
• training=False
• mask=None
```
|
non_process
|
can t excute convnext with input tensor when keras applications convnext is executed with input tensor it is not executed the cause seems to be a problem that occurs as it changes to the form of when passing utils layer utils get source inputs when using input shape there is no problem this problem was identified in versions tensorflow and keras check please the following error message valueerror traceback most recent call last input in in x tf keras layers input shape model tf keras applications convnexttiny input tensor x include preprocessing true include top false weights none file opt lib site packages keras applications convnext py in convnexttiny model name include top include preprocessing weights input tensor input shape pooling classes classifier activation keras export keras applications convnext convnexttiny keras applications convnexttiny classifier activation softmax return convnext depths model configs projection dims model configs drop path rate layer scale init value default size model configs model name model name include top include top include preprocessing include preprocessing weights weights input tensor input tensor input shape input shape pooling pooling classes classes classifier activation classifier activation file opt lib site packages keras applications convnext py in convnext depths projection dims drop path rate layer scale init value default size model name include preprocessing include top weights input tensor input shape pooling classes classifier activation num channels input shape if num channels x prestem name model name x stem block stem sequential sequential layers name model name stem file opt lib site packages keras applications convnext py in prestem apply x def apply x x layers normalization mean variance name name prestem normalization x return x file opt lib site packages keras utils traceback utils py in filter traceback error handler args kwargs filtered tb process traceback frames e traceback to get the full stack trace call tf debugging disable traceback filtering raise e with traceback filtered tb from none finally del filtered tb file opt lib site packages keras layers preprocessing normalization py in normalization build self input shape super build input shape if isinstance input shape list tuple and all isinstance shape tf tensorshape for shape in input shape raise valueerror normalization only accepts a single input if you are passing a python list or tuple as a single input please convert to a numpy array or tf tensor input shape tf tensorshape input shape as list ndim len input shape valueerror normalization only accepts a single input if you are passing a python list or tuple as a single input please convert to a numpy array or tf tensor the following error message without include preprocessing warning tensorflow layers in a sequential model should only have a single input tensor received inputs consider rewriting this model with the functional api typeerror traceback most recent call last input in in x tf keras layers input shape model tf keras applications convnexttiny input tensor x include preprocessing false include top false weights none file opt lib site packages keras applications convnext py in convnexttiny model name include top include preprocessing weights input tensor input shape pooling classes classifier activation keras export keras applications convnext convnexttiny keras applications convnexttiny classifier activation softmax return convnext depths model configs projection dims model configs drop path rate layer scale init value default size model configs model name model name include top include top include preprocessing include preprocessing weights weights input tensor input tensor input shape input shape pooling pooling classes classes classifier activation classifier activation file opt lib site packages keras applications convnext py in convnext depths projection dims drop path rate layer scale init value default size model name include preprocessing include top weights input tensor input shape pooling classes classifier activation num convnext blocks for i in range num convnext blocks x downsample layers x for j in range depths x convnextblock projection dim projection dims drop path rate depth drop rates layer scale init value layer scale init value name model name f stage i block j x file opt lib site packages keras utils traceback utils py in filter traceback error handler args kwargs filtered tb process traceback frames e traceback to get the full stack trace call tf debugging disable traceback filtering raise e with traceback filtered tb from none finally del filtered tb file in raise from value from value typeerror exception encountered when calling layer convnext tiny stem type sequential dimension value must be integer or none or have an index method got value tensorshape with type call arguments received by layer convnext tiny stem type sequential • inputs • training false • mask none
| 0
|
21,413
| 29,359,588,912
|
IssuesEvent
|
2023-05-28 00:36:15
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] Node.js Developer na Coodesh
|
SALVADOR BACK-END INFRAESTRUTURA FULL-STACK SCRUM BDD GIT TYPESCRIPT NODE.JS DOCKER DEVOPS REACT AWS REMOTO PROCESSOS INOVAÇÃO BACKEND GITHUB KANBAN CI CD SEGURANÇA GITFLOW UMA C QUALIDADE CLEAN XP TESTES AUTOMATIZADOS MICROSERVICES METODOLOGIAS ÁGEIS EXPRESS NEGÓCIOS MONITORAMENTO SRE PAAS Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>O Grupo fácil está buscando <strong><ins>Node Developer</ins></strong> para fazer parte de seu time!</p>
<p></p>
<p><strong>🏢 Quem somos: </strong></p>
<p>O Grupo Fácil está focado em desenvolvimento de produtos na área da saúde, com o objetivo principal em atender os grandes nomes em operadoras de saúde do mercado brasileiro.</p>
<p>Há mais de 27 anos dedicados na entrega de valor aos nossos clientes, focados na melhoria contínua e evolução natural no mercado de software brasileiro.</p>
<p>Atualmente ingressamos em uma nova grande jornada de inovação, onde novos processos e produtos serão elaborados para entregar mais valor a nossos clientes, por isso, esperamos que você faça parte dessa jornada conosco. </p>
<p></p>
<p><strong>💻 Como é nosso time? </strong></p>
<p>Trabalhamos de forma remota, usando as práticas de comunicação assíncrona e workflow para o desenvolvimento ágil, a principal ideia é trabalhar duro, focados nas metas e objetivos, porém sem perder a liberdade e a inspiração que nos faz acordar cedo todos os dias.</p>
<p>Temos pessoas focadas no negócio, no frontend, outros no backend e até no perfil full-stack, esperamos de você que nos mostre onde se sente melhor. 😉 </p>
<p></p>
<p><strong>O que esperamos de você?</strong></p>
<ul>
<li>Transparência (comunicação clara e eficiente sem ruído)</li>
<li>Senso crítico com as demandas e processos</li>
<li>Experiência com desenvolvimento de grandes projeto (escala, performance, qualidade e etc)</li>
<li>Experiência na codificação de testes automatizados (unit test, BDD, Integration test)</li>
<li>Experiência com Git/Github e GitFlow (já usou github actions?!)</li>
<li>Capacidade de pedir ajuda e/ou ajudar em momentos de crise/impedimentos</li>
<li>Capacidade de "se virar", "correr atrás" de algo desconhecido, assumir riscos (com parcimônia)</li>
<li>Capacidade de ensinar, instruir e orientar pessoas menos experientes do projeto ou time (multiplique o conhecimento)</li>
</ul>
<p></p>
<p><strong>Vamos falar de código? </strong></p>
<ul>
<li>Aqui usamos Typescript com S.O.L.I.D</li>
<li>Usamos microservices, docker, AWS e tudo que há de bom 😁</li>
<li>A arquitetura é desacoplada, testada com Hexagonal (Ports and Adapters)</li>
<li>No frontend usamos React, Styled-Components, Vite, Storybook e etc</li>
<li>No backend usamos Node.js, Express/Fastify, TypeORM, Oracle e etc</li>
<li>E mais alguns detalhes que ficaremos felizes em contar durante nossa entrevista técnica.</li>
<li>E ai, se sente confortável com isso?</li>
</ul>
<p></p>
## Grupo Fácil:
<p>Ao longo de 27 anos de história, o Grupo Fácil se tornou referência nacional em sistemas, softwares e serviços para a gestão de negócios nas áreas financeira e de crédito, da saúde e no setor imobiliário.</p>
<p>O Grupo Fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes. </p><a href='https://coodesh.com/empresas/grupo-facil'>Veja mais no site</a>
## Habilidades:
- Node.js
- Express.js
- TypeORM
- Oracle
## Local:
100% Remoto
## Diferenciais:
- Conhecimento de Clean Architecture e/ou Hexagonal Architecture;
- Conhecimentos/Experiência de automação e/ou infraestrutura como código (DevOps, CI/CD);
- Conhecimento de metodologias ágeis (Kanban, Scrum, XP, Scrumban);
- Conhecimento e/ou certificação em PaaS/Clouds (Preferencialmente AWS);
- Fila, processamento assíncrono, tópicos, modelo pub/sub;
- Sabe algo sobre monitoramento, autoscalling, stress-test, load-test, SRE??
## Benefícios:
- Convênio com farmácia;
- Participação nos lucros;
- Vale refeição;
- Vale transporte;
- Parcerias e convênios;
- Programas de saúde e bem-estar.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Node.js Developer na Grupo Fácil](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Back-End
|
1.0
|
[Remoto] Node.js Developer na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>O Grupo fácil está buscando <strong><ins>Node Developer</ins></strong> para fazer parte de seu time!</p>
<p></p>
<p><strong>🏢 Quem somos: </strong></p>
<p>O Grupo Fácil está focado em desenvolvimento de produtos na área da saúde, com o objetivo principal em atender os grandes nomes em operadoras de saúde do mercado brasileiro.</p>
<p>Há mais de 27 anos dedicados na entrega de valor aos nossos clientes, focados na melhoria contínua e evolução natural no mercado de software brasileiro.</p>
<p>Atualmente ingressamos em uma nova grande jornada de inovação, onde novos processos e produtos serão elaborados para entregar mais valor a nossos clientes, por isso, esperamos que você faça parte dessa jornada conosco. </p>
<p></p>
<p><strong>💻 Como é nosso time? </strong></p>
<p>Trabalhamos de forma remota, usando as práticas de comunicação assíncrona e workflow para o desenvolvimento ágil, a principal ideia é trabalhar duro, focados nas metas e objetivos, porém sem perder a liberdade e a inspiração que nos faz acordar cedo todos os dias.</p>
<p>Temos pessoas focadas no negócio, no frontend, outros no backend e até no perfil full-stack, esperamos de você que nos mostre onde se sente melhor. 😉 </p>
<p></p>
<p><strong>O que esperamos de você?</strong></p>
<ul>
<li>Transparência (comunicação clara e eficiente sem ruído)</li>
<li>Senso crítico com as demandas e processos</li>
<li>Experiência com desenvolvimento de grandes projeto (escala, performance, qualidade e etc)</li>
<li>Experiência na codificação de testes automatizados (unit test, BDD, Integration test)</li>
<li>Experiência com Git/Github e GitFlow (já usou github actions?!)</li>
<li>Capacidade de pedir ajuda e/ou ajudar em momentos de crise/impedimentos</li>
<li>Capacidade de "se virar", "correr atrás" de algo desconhecido, assumir riscos (com parcimônia)</li>
<li>Capacidade de ensinar, instruir e orientar pessoas menos experientes do projeto ou time (multiplique o conhecimento)</li>
</ul>
<p></p>
<p><strong>Vamos falar de código? </strong></p>
<ul>
<li>Aqui usamos Typescript com S.O.L.I.D</li>
<li>Usamos microservices, docker, AWS e tudo que há de bom 😁</li>
<li>A arquitetura é desacoplada, testada com Hexagonal (Ports and Adapters)</li>
<li>No frontend usamos React, Styled-Components, Vite, Storybook e etc</li>
<li>No backend usamos Node.js, Express/Fastify, TypeORM, Oracle e etc</li>
<li>E mais alguns detalhes que ficaremos felizes em contar durante nossa entrevista técnica.</li>
<li>E ai, se sente confortável com isso?</li>
</ul>
<p></p>
## Grupo Fácil:
<p>Ao longo de 27 anos de história, o Grupo Fácil se tornou referência nacional em sistemas, softwares e serviços para a gestão de negócios nas áreas financeira e de crédito, da saúde e no setor imobiliário.</p>
<p>O Grupo Fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes. </p><a href='https://coodesh.com/empresas/grupo-facil'>Veja mais no site</a>
## Habilidades:
- Node.js
- Express.js
- TypeORM
- Oracle
## Local:
100% Remoto
## Diferenciais:
- Conhecimento de Clean Architecture e/ou Hexagonal Architecture;
- Conhecimentos/Experiência de automação e/ou infraestrutura como código (DevOps, CI/CD);
- Conhecimento de metodologias ágeis (Kanban, Scrum, XP, Scrumban);
- Conhecimento e/ou certificação em PaaS/Clouds (Preferencialmente AWS);
- Fila, processamento assíncrono, tópicos, modelo pub/sub;
- Sabe algo sobre monitoramento, autoscalling, stress-test, load-test, SRE??
## Benefícios:
- Convênio com farmácia;
- Participação nos lucros;
- Vale refeição;
- Vale transporte;
- Parcerias e convênios;
- Programas de saúde e bem-estar.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Node.js Developer na Grupo Fácil](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Back-End
|
process
|
node js developer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 o grupo fácil está buscando node developer para fazer parte de seu time 🏢 quem somos o grupo fácil está focado em desenvolvimento de produtos na área da saúde com o objetivo principal em atender os grandes nomes em operadoras de saúde do mercado brasileiro há mais de anos dedicados na entrega de valor aos nossos clientes focados na melhoria contínua e evolução natural no mercado de software brasileiro atualmente ingressamos em uma nova grande jornada de inovação onde novos processos e produtos serão elaborados para entregar mais valor a nossos clientes por isso esperamos que você faça parte dessa jornada conosco nbsp 💻 como é nosso time trabalhamos de forma remota usando as práticas de comunicação assíncrona e workflow para o desenvolvimento ágil a principal ideia é trabalhar duro focados nas metas e objetivos porém sem perder a liberdade e a inspiração que nos faz acordar cedo todos os dias temos pessoas focadas no negócio no frontend outros no backend e até no perfil full stack esperamos de você que nos mostre onde se sente melhor 😉 nbsp o que esperamos de você transparência comunicação clara e eficiente sem ruído senso crítico com as demandas e processos experiência com desenvolvimento de grandes projeto escala performance qualidade e etc experiência na codificação de testes automatizados unit test bdd integration test experiência com git github e gitflow já usou github actions capacidade de pedir ajuda e ou ajudar em momentos de crise impedimentos capacidade de se virar correr atrás de algo desconhecido assumir riscos com parcimônia capacidade de ensinar instruir e orientar pessoas menos experientes do projeto ou time multiplique o conhecimento vamos falar de código aqui usamos typescript com s o l i d usamos microservices docker aws e tudo que há de bom 😁 a arquitetura é desacoplada testada com hexagonal ports and adapters no frontend usamos react styled components vite storybook e etc no backend usamos node js express fastify typeorm oracle e etc e mais alguns detalhes que ficaremos felizes em contar durante nossa entrevista técnica e ai se sente confortável com isso grupo fácil ao longo de anos de história o grupo fácil se tornou referência nacional em sistemas softwares e serviços para a gestão de negócios nas áreas financeira e de crédito da saúde e no setor imobiliário o grupo fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes nbsp habilidades node js express js typeorm oracle local remoto diferenciais conhecimento de clean architecture e ou hexagonal architecture conhecimentos experiência de automação e ou infraestrutura como código devops ci cd conhecimento de metodologias ágeis kanban scrum xp scrumban conhecimento e ou certificação em paas clouds preferencialmente aws fila processamento assíncrono tópicos modelo pub sub sabe algo sobre monitoramento autoscalling stress test load test sre benefícios convênio com farmácia participação nos lucros vale refeição vale transporte parcerias e convênios programas de saúde e bem estar como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria back end
| 1
|
11,876
| 14,675,535,217
|
IssuesEvent
|
2020-12-30 17:50:40
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
gdal polygonize fails with "No module named '_gdal'. message
|
Bug Feedback Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
Running Raster -> Conversion -> Polygonize fails. The log contains the exception ModuleNotFoundError with message "No module named 'osgeo._gdal'"
Handling this exception produces another with message "No module named '_gdal'.
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
Load a raster.
Use menu Raster -> Conversion -> Polygonize
Choose raster as input layer and click Run
**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
The error appears in 3.10.13. It is not in 3.10.8.
QGIS version | 3.10.13-A Coruña | QGIS code revision | 311f3bbade
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | plugin_reloader; QSWAT3; QSWAT3_64; QSWATPlus3_64; remotedebug; db_manager; MetaSearch; processing
**Additional context**
<!-- Add any other context about the problem here. -->
|
1.0
|
gdal polygonize fails with "No module named '_gdal'. message - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
Running Raster -> Conversion -> Polygonize fails. The log contains the exception ModuleNotFoundError with message "No module named 'osgeo._gdal'"
Handling this exception produces another with message "No module named '_gdal'.
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
Load a raster.
Use menu Raster -> Conversion -> Polygonize
Choose raster as input layer and click Run
**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
The error appears in 3.10.13. It is not in 3.10.8.
QGIS version | 3.10.13-A Coruña | QGIS code revision | 311f3bbade
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | plugin_reloader; QSWAT3; QSWAT3_64; QSWATPlus3_64; remotedebug; db_manager; MetaSearch; processing
**Additional context**
<!-- Add any other context about the problem here. -->
|
process
|
gdal polygonize fails with no module named gdal message bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug running raster conversion polygonize fails the log contains the exception modulenotfounderror with message no module named osgeo gdal handling this exception produces another with message no module named gdal how to reproduce load a raster use menu raster conversion polygonize choose raster as input layer and click run qgis and os versions about click in the table ctrl a and then ctrl c finally paste here the error appears in it is not in qgis version a coruña qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins plugin reloader remotedebug db manager metasearch processing additional context
| 1
|
183,348
| 6,687,954,207
|
IssuesEvent
|
2017-10-08 08:18:02
|
CruCentralCoast/CruiOS
|
https://api.github.com/repos/CruCentralCoast/CruiOS
|
closed
|
About Section: Google Form Link is Editable
|
Priority: Minor Type: Bug
|
On the About section, we have a link to a Google Form for developer feedback. Instead of going to the link, the app opens a keyboard and allows you to type more letters and numbers onto the URL. I have no idea if this affected the data on the server, so if it did, the Google Form link will have to be changed back to what it was. The URL should act as a URL and open the form.
iOS 11.0.2
Version 1.4 (20)
iPhone 6
|
1.0
|
About Section: Google Form Link is Editable - On the About section, we have a link to a Google Form for developer feedback. Instead of going to the link, the app opens a keyboard and allows you to type more letters and numbers onto the URL. I have no idea if this affected the data on the server, so if it did, the Google Form link will have to be changed back to what it was. The URL should act as a URL and open the form.
iOS 11.0.2
Version 1.4 (20)
iPhone 6
|
non_process
|
about section google form link is editable on the about section we have a link to a google form for developer feedback instead of going to the link the app opens a keyboard and allows you to type more letters and numbers onto the url i have no idea if this affected the data on the server so if it did the google form link will have to be changed back to what it was the url should act as a url and open the form ios version iphone
| 0
|
319,630
| 27,389,839,343
|
IssuesEvent
|
2023-02-28 15:36:05
|
USEPA/haztrak
|
https://api.github.com/repos/USEPA/haztrak
|
opened
|
Quicker SIgn button and form
|
feature reactjs test
|
# 🚀 Feature Request
Add a button to the manifest view to submit a Quicker Sign signature to e-Manifest.
The Button should be disabled when...
- The manifest is editable. the `readOnly` prop should be equal to `true`.
- The user's RcraProfile.is_api_user field is `False`
- The manifest status signifies that it is not time for the site, who the user is currently viewing as, to sign (awaiting signature of earlier handler, already been signed, etc)
## Additional to do items
- This information is not currently stored in the Redux store (see new issue)
- This information is not currently serialized to JSON when requested
- We need to expand on this property to account for groups to allow admins to add non-API credential holders to use the RCRAInfo web services using another user's API credentials.
<!--
Please provide a clear and concise description of what you'd like to see. Things like...
1. what is the expected behavior?
2. your opinion on what the default(s) should be?
-->
|
1.0
|
Quicker SIgn button and form - # 🚀 Feature Request
Add a button to the manifest view to submit a Quicker Sign signature to e-Manifest.
The Button should be disabled when...
- The manifest is editable. the `readOnly` prop should be equal to `true`.
- The user's RcraProfile.is_api_user field is `False`
- The manifest status signifies that it is not time for the site, who the user is currently viewing as, to sign (awaiting signature of earlier handler, already been signed, etc)
## Additional to do items
- This information is not currently stored in the Redux store (see new issue)
- This information is not currently serialized to JSON when requested
- We need to expand on this property to account for groups to allow admins to add non-API credential holders to use the RCRAInfo web services using another user's API credentials.
<!--
Please provide a clear and concise description of what you'd like to see. Things like...
1. what is the expected behavior?
2. your opinion on what the default(s) should be?
-->
|
non_process
|
quicker sign button and form 🚀 feature request add a button to the manifest view to submit a quicker sign signature to e manifest the button should be disabled when the manifest is editable the readonly prop should be equal to true the user s rcraprofile is api user field is false the manifest status signifies that it is not time for the site who the user is currently viewing as to sign awaiting signature of earlier handler already been signed etc additional to do items this information is not currently stored in the redux store see new issue this information is not currently serialized to json when requested we need to expand on this property to account for groups to allow admins to add non api credential holders to use the rcrainfo web services using another user s api credentials please provide a clear and concise description of what you d like to see things like what is the expected behavior your opinion on what the default s should be
| 0
|
1,028
| 3,487,681,885
|
IssuesEvent
|
2016-01-02 05:28:21
|
spootTheLousy/saguaro
|
https://api.github.com/repos/spootTheLousy/saguaro
|
opened
|
Deletion class doesn't check for ghost bumping.
|
Administrative Post/text processing REVISIT
|
TO-DO:
Prevent users from ghost bumping. Deletion timers maybe?
|
1.0
|
Deletion class doesn't check for ghost bumping. - TO-DO:
Prevent users from ghost bumping. Deletion timers maybe?
|
process
|
deletion class doesn t check for ghost bumping to do prevent users from ghost bumping deletion timers maybe
| 1
|
15,981
| 20,188,186,289
|
IssuesEvent
|
2022-02-11 01:16:17
|
savitamittalmsft/WAS-SEC-TEST
|
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
|
opened
|
Protect all public endpoints with appropriate controls
|
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Networking & Connectivity Endpoints
|
<a href="https://docs.microsoft.com/azure/architecture/framework/security/design-network-endpoints#web-application-firewalls-wafs">Protect all public endpoints with appropriate controls</a>
<p><b>Why Consider This?</b></p>
External application endpoints should be protected against common attack vectors, from Denial of Service (DoS) attacks like Slowloris to app-level exploits, to prevent potential application downtime due to malicious intent. Azure-native technologies such as Azure Firewall, Application Gateway/Azure Front Door, WAF, and DDoS Protection Standard Plan can be used to achieve requisite protection (Azure DDoS Protection).
<p><b>Context</b></p>
<p><b>Suggested Actions</b></p>
<p><span>Protect all public endpoints with appropriate solutions, e.g. Azure Front Door, Application Gateway, Azure Firewall, Azure DDOS Protection or any 3rd party solution.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/firewall/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/firewall/overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/frontdoor/" target="_blank"><span>https://docs.microsoft.com/en-us/azure/frontdoor/</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/application-gateway/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/application-gateway/overview</span></a><span /></p>
|
1.0
|
Protect all public endpoints with appropriate controls - <a href="https://docs.microsoft.com/azure/architecture/framework/security/design-network-endpoints#web-application-firewalls-wafs">Protect all public endpoints with appropriate controls</a>
<p><b>Why Consider This?</b></p>
External application endpoints should be protected against common attack vectors, from Denial of Service (DoS) attacks like Slowloris to app-level exploits, to prevent potential application downtime due to malicious intent. Azure-native technologies such as Azure Firewall, Application Gateway/Azure Front Door, WAF, and DDoS Protection Standard Plan can be used to achieve requisite protection (Azure DDoS Protection).
<p><b>Context</b></p>
<p><b>Suggested Actions</b></p>
<p><span>Protect all public endpoints with appropriate solutions, e.g. Azure Front Door, Application Gateway, Azure Firewall, Azure DDOS Protection or any 3rd party solution.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/firewall/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/firewall/overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/frontdoor/" target="_blank"><span>https://docs.microsoft.com/en-us/azure/frontdoor/</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/application-gateway/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/application-gateway/overview</span></a><span /></p>
|
process
|
protect all public endpoints with appropriate controls why consider this external application endpoints should be protected against common attack vectors from denial of service dos attacks like slowloris to app level exploits to prevent potential application downtime due to malicious intent azure native technologies such as azure firewall application gateway azure front door waf and ddos protection standard plan can be used to achieve requisite protection azure ddos protection context suggested actions protect all public endpoints with appropriate solutions e g azure front door application gateway azure firewall azure ddos protection or any party solution learn more
| 1
|
2,908
| 5,891,937,182
|
IssuesEvent
|
2017-05-17 18:16:21
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process: closes before flushing all data
|
child_process
|
I use NodeJS v5.1.0 x64 on Windows 10. The following code
``` javascript
var cp = require('child_process');
var stdout = "";
var stderr = "";
var process = cp.spawn("git", ["ls-remote", "--tags", "--heads", "git://github.com/jquery/jquery.git"]);
process.stdout.on('data', function (data) {
var text = data.toString();
stdout += text;
});
process.stderr.on('data', function (data) {
var text = data.toString();
stderr += text;
});
process.on('error', function (error) {
debugger;
});
process.on('close', function (code) {
debugger;
});
process.on('exit', function (code) {
debugger;
});
```
results in the following text in stdout and then the `exit` and `close` events are triggered. You can see that the response is incomplete. When I call the command in the Windows CMD directly, I get a complete response back.
```
ff619f3b978304700add92952485a5fb0a8fd80c refs/heads/1.12-stable
4fea389ad2b7515bc3a9081245979ab7be566b99 refs/heads/2.2-stable
83391859bda90f630bf02a5e04d82e9f1babeb1f refs/heads/compat
90347a78d69b15a515c462eb378adb7e6067f5dd refs/heads/killphp
eaa3e9f0cfc68083556cf61195821d90e369f646 refs/heads/master
bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0
bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0.1
dbee06de85859af59591813d3004e6695b8bb278 refs/tags/1.0.2
4e3da33c59fafe34e237585743e86e24ba81046e refs/tags/1.0.3
ed37ca66e5f79b4250f38cbd7a7f809ebfc15a20 refs/tags/1.0.4
73f55becc5e056ad28f6aec22ec014ffb3ab913b refs/tags/1.0a
2aa67026ebe6bea90fd137fc99b4c9422977e3f0 refs/tags/1.1
2550b1ea6a6eb6ba50d378a5dcfbc8146ca88bff refs/tags/1.1.1
92973279ea26c415f505d187276460115c668f17 refs/tags/1.1.2
fdd437800d983346c5936abedcf292c2a612b892 refs/tags/1.1.3
f70010b0ad13e3e4d6366287fd79636a9ceca1cd refs/tags/1.1.3.1
19f5cf0c893aa2ba2ffb00c00c7efc1db4383f3e refs/tags/1.1.3a
557ec9a92f40f97254e48240825fcb9be5a08e20 refs/tags/1.1.4
2c62b787b2cafb32af14edadc963b8ba3cb847b8 refs/tags/1.10.0
3a8fcde250da0db237396e9ab97fb099eb68c624 refs/tags/1.10.0-beta1
935eb19d87063ec30b717b3d45368d3c73cb591d refs/tags/1.10.1
16b079b164d62bd807c612806842a13bf9b04d17 refs/tags/1.10.2
91cd1587ed4a1d4f3834227e8eb3cc30cb385409 refs/tags/1.11.0
b44595824c3f23441eab2f34551587c5ae6fd31c refs/tags/1.11.0-beta1
02cb5be0bb5a49cda3a187d169d89d91874bbf1e refs/tags/1.11.0-beta2
5bb568cd61cb127ef4ebcc63bb48a37cd5af562d refs/tags/1.11.0-beta3
5c62f3cd019218e1b401ce2cf4cad644772e387f refs/tags/1.11.0-rc1
0d5ec2d8ac94a419ee47a39319c43ff9a7326b50 refs/tags/1.11.1
b6c4e50fb8f24dcf7bf38e053e68c4d0e7f059e8 refs/tags/1.11.1-beta1
baa21ff7a91ca165b63a5f900313da14c36fe255 refs/tags/1.11.1-rc1
b50c9a8d92f94f310138f4528e3a36b5336539a1 refs/tags/1.11.1-rc2
9690801db01709bfbff5f977d07fb7cc14472908 refs/tags/1.11.2
1472290917f17af05e98007136096784f9051fab refs/tags/1.11.3
8bcb016dfaa029a5e98cc3c25b9f999053217f36 refs/tags/1.1a
d1390665b6e7d475d237cafcf8948440c9adfbf8 refs/tags/1.1b
ce256312d518e16a6ee7df5f7f301b8bf9a1caf2 refs/tags/1.2
456da2d4a805012003df3c1b173148368d3536f7 refs/tags/1.2.1
6c81082a6b20880cbd5d35ec5c50d4270f3af2ce refs/tags/1.2.2
13a5fdb6a1571012c3e065a991fb4eff083057fb refs/tags/1.2.2b
3f839a69c5682419c921a1a2110289c8941556b0 refs/tags/1.2.2b2
d9a3c4fefaf84db7eba9b1f8204d1e134266b823 refs/tags/1.2.3
32ba00081ddd56c00655d0a7b7825ac947247cc6 refs/tags/1.2.3a
2c07690377522119891784ec11c89e87f0a12545 refs/tags/1.2.3b
2db27efd662693dd90fbf89ea1e4bbee7c9fb929 refs/tags/1.2.4
1ba2865b089fc06242004b7003b4f8dcfd92dc13 refs/tags/1.2.4a
44a6f5d0cb7f1124cd80e9b72d1fff9a46fe4c0b refs/tags/1.2.4b
456d8a62ccf3244c58ace2b315694ad54cfc05e3 refs/tags/1.2.5
ee931edfea95a0ed49aa4382b5c09cae6f7dafdb refs/tags/1.2.6
9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3
9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3.0
6df57d1b82c88ea85d2be081ee8f2362e1c362b7 refs/tags/1.3.1
8b79ea13656727e2449807a812dd0fd8c2783aad refs/tags/1.3.1rc1
13e9cde0efeadbbc5b3a375ee642bed394dd3a59 refs/tags/1.3.2
df1f656aae019bab1c0e84b41aca26714b4d0d5c refs/tags/1.3b1
c786f1097ee26095f2f7eb721a450c550775ba84 refs/tags/1.3b2
869afc46cd8229f41d82380f01830dd59a0084f7 refs/tags/1.3rc1
4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4
4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4.0
15d261b9673e153f9e9710da897de62c306afe53 refs/tags/1.4.1
9e4beb5ac258bfc834913c1781891ef49b2deb93 refs/tags/1.4.2
07458abcaf02b5e991223953081bb0dd5240e2c1 refs/tags/1.4.3
66a788ab837fd9b483740ee0b44c1e76f846aadd refs/tags/1.4.3rc1
3e5520fbdc7231b3f38e145020b40524c1e6654d refs/tags/1.4.3rc2
9b20df0375f37cc0997bd8ed3d1892a23967d3d5 refs/tags/1.4.4
c22623602ee3c9ff70df9e6f51132931d19bab04 refs/tags/1.4.4rc1
4d7f7453dca8e385cfb02089a8d5b2c47371bba3 refs/tags/1.4.4rc2
50708c070704c607a0ddc51679b93845c8fee5e1 refs/tags/1.4.4rc3
c90a1cb61b41a5686bc3764f144a156e734ed459 refs/tags/1.4a1
f745357baec82e0be59c651f70d1e489f237c398 refs/tags/1.4a2
857d8afd6917775b7cef6fd22e57dd9acec77f01 refs/tags/1.4rc1
675226eea749
```
EDIT:
The same happens on `NodeJS v4.2.2 x86`
|
1.0
|
child_process: closes before flushing all data - I use NodeJS v5.1.0 x64 on Windows 10. The following code
``` javascript
var cp = require('child_process');
var stdout = "";
var stderr = "";
var process = cp.spawn("git", ["ls-remote", "--tags", "--heads", "git://github.com/jquery/jquery.git"]);
process.stdout.on('data', function (data) {
var text = data.toString();
stdout += text;
});
process.stderr.on('data', function (data) {
var text = data.toString();
stderr += text;
});
process.on('error', function (error) {
debugger;
});
process.on('close', function (code) {
debugger;
});
process.on('exit', function (code) {
debugger;
});
```
results in the following text in stdout and then the `exit` and `close` events are triggered. You can see that the response is incomplete. When I call the command in the Windows CMD directly, I get a complete response back.
```
ff619f3b978304700add92952485a5fb0a8fd80c refs/heads/1.12-stable
4fea389ad2b7515bc3a9081245979ab7be566b99 refs/heads/2.2-stable
83391859bda90f630bf02a5e04d82e9f1babeb1f refs/heads/compat
90347a78d69b15a515c462eb378adb7e6067f5dd refs/heads/killphp
eaa3e9f0cfc68083556cf61195821d90e369f646 refs/heads/master
bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0
bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0.1
dbee06de85859af59591813d3004e6695b8bb278 refs/tags/1.0.2
4e3da33c59fafe34e237585743e86e24ba81046e refs/tags/1.0.3
ed37ca66e5f79b4250f38cbd7a7f809ebfc15a20 refs/tags/1.0.4
73f55becc5e056ad28f6aec22ec014ffb3ab913b refs/tags/1.0a
2aa67026ebe6bea90fd137fc99b4c9422977e3f0 refs/tags/1.1
2550b1ea6a6eb6ba50d378a5dcfbc8146ca88bff refs/tags/1.1.1
92973279ea26c415f505d187276460115c668f17 refs/tags/1.1.2
fdd437800d983346c5936abedcf292c2a612b892 refs/tags/1.1.3
f70010b0ad13e3e4d6366287fd79636a9ceca1cd refs/tags/1.1.3.1
19f5cf0c893aa2ba2ffb00c00c7efc1db4383f3e refs/tags/1.1.3a
557ec9a92f40f97254e48240825fcb9be5a08e20 refs/tags/1.1.4
2c62b787b2cafb32af14edadc963b8ba3cb847b8 refs/tags/1.10.0
3a8fcde250da0db237396e9ab97fb099eb68c624 refs/tags/1.10.0-beta1
935eb19d87063ec30b717b3d45368d3c73cb591d refs/tags/1.10.1
16b079b164d62bd807c612806842a13bf9b04d17 refs/tags/1.10.2
91cd1587ed4a1d4f3834227e8eb3cc30cb385409 refs/tags/1.11.0
b44595824c3f23441eab2f34551587c5ae6fd31c refs/tags/1.11.0-beta1
02cb5be0bb5a49cda3a187d169d89d91874bbf1e refs/tags/1.11.0-beta2
5bb568cd61cb127ef4ebcc63bb48a37cd5af562d refs/tags/1.11.0-beta3
5c62f3cd019218e1b401ce2cf4cad644772e387f refs/tags/1.11.0-rc1
0d5ec2d8ac94a419ee47a39319c43ff9a7326b50 refs/tags/1.11.1
b6c4e50fb8f24dcf7bf38e053e68c4d0e7f059e8 refs/tags/1.11.1-beta1
baa21ff7a91ca165b63a5f900313da14c36fe255 refs/tags/1.11.1-rc1
b50c9a8d92f94f310138f4528e3a36b5336539a1 refs/tags/1.11.1-rc2
9690801db01709bfbff5f977d07fb7cc14472908 refs/tags/1.11.2
1472290917f17af05e98007136096784f9051fab refs/tags/1.11.3
8bcb016dfaa029a5e98cc3c25b9f999053217f36 refs/tags/1.1a
d1390665b6e7d475d237cafcf8948440c9adfbf8 refs/tags/1.1b
ce256312d518e16a6ee7df5f7f301b8bf9a1caf2 refs/tags/1.2
456da2d4a805012003df3c1b173148368d3536f7 refs/tags/1.2.1
6c81082a6b20880cbd5d35ec5c50d4270f3af2ce refs/tags/1.2.2
13a5fdb6a1571012c3e065a991fb4eff083057fb refs/tags/1.2.2b
3f839a69c5682419c921a1a2110289c8941556b0 refs/tags/1.2.2b2
d9a3c4fefaf84db7eba9b1f8204d1e134266b823 refs/tags/1.2.3
32ba00081ddd56c00655d0a7b7825ac947247cc6 refs/tags/1.2.3a
2c07690377522119891784ec11c89e87f0a12545 refs/tags/1.2.3b
2db27efd662693dd90fbf89ea1e4bbee7c9fb929 refs/tags/1.2.4
1ba2865b089fc06242004b7003b4f8dcfd92dc13 refs/tags/1.2.4a
44a6f5d0cb7f1124cd80e9b72d1fff9a46fe4c0b refs/tags/1.2.4b
456d8a62ccf3244c58ace2b315694ad54cfc05e3 refs/tags/1.2.5
ee931edfea95a0ed49aa4382b5c09cae6f7dafdb refs/tags/1.2.6
9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3
9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3.0
6df57d1b82c88ea85d2be081ee8f2362e1c362b7 refs/tags/1.3.1
8b79ea13656727e2449807a812dd0fd8c2783aad refs/tags/1.3.1rc1
13e9cde0efeadbbc5b3a375ee642bed394dd3a59 refs/tags/1.3.2
df1f656aae019bab1c0e84b41aca26714b4d0d5c refs/tags/1.3b1
c786f1097ee26095f2f7eb721a450c550775ba84 refs/tags/1.3b2
869afc46cd8229f41d82380f01830dd59a0084f7 refs/tags/1.3rc1
4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4
4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4.0
15d261b9673e153f9e9710da897de62c306afe53 refs/tags/1.4.1
9e4beb5ac258bfc834913c1781891ef49b2deb93 refs/tags/1.4.2
07458abcaf02b5e991223953081bb0dd5240e2c1 refs/tags/1.4.3
66a788ab837fd9b483740ee0b44c1e76f846aadd refs/tags/1.4.3rc1
3e5520fbdc7231b3f38e145020b40524c1e6654d refs/tags/1.4.3rc2
9b20df0375f37cc0997bd8ed3d1892a23967d3d5 refs/tags/1.4.4
c22623602ee3c9ff70df9e6f51132931d19bab04 refs/tags/1.4.4rc1
4d7f7453dca8e385cfb02089a8d5b2c47371bba3 refs/tags/1.4.4rc2
50708c070704c607a0ddc51679b93845c8fee5e1 refs/tags/1.4.4rc3
c90a1cb61b41a5686bc3764f144a156e734ed459 refs/tags/1.4a1
f745357baec82e0be59c651f70d1e489f237c398 refs/tags/1.4a2
857d8afd6917775b7cef6fd22e57dd9acec77f01 refs/tags/1.4rc1
675226eea749
```
EDIT:
The same happens on `NodeJS v4.2.2 x86`
|
process
|
child process closes before flushing all data i use nodejs on windows the following code javascript var cp require child process var stdout var stderr var process cp spawn git process stdout on data function data var text data tostring stdout text process stderr on data function data var text data tostring stderr text process on error function error debugger process on close function code debugger process on exit function code debugger results in the following text in stdout and then the exit and close events are triggered you can see that the response is incomplete when i call the command in the windows cmd directly i get a complete response back refs heads stable refs heads stable refs heads compat refs heads killphp refs heads master refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags edit the same happens on nodejs
| 1
|
738,390
| 25,556,503,700
|
IssuesEvent
|
2022-11-30 07:14:32
|
pc2ccs/pc2v9
|
https://api.github.com/repos/pc2ccs/pc2v9
|
closed
|
event-feed API endpoint returns incorrect ordinal field for "runs" testcases
|
bug high priority CLICS CCS Update
|
**Describe the issue**:
Reported by DOMJudge: For `"runs" `test cases in the event-feed, the `"ordinal"` value should start at 1 not 0 as per the CLICS Specification. Currently, the first test case has an ordinal value of `"0"`.
**To Reproduce**:
Start up a contest that has some judged runs in it.
Start the Event Feed.
Connect to the event-feed endpoint and download the event feed.
Note that the first testcase for each submission has an `"ordinal"` value of 0. Consequently, all following test cases are off by one for `"ordinal"`
**Expected behavior**:
The `"ordinal" `value of the first test case should start at 1, not 0.
**Actual behavior**:
The `"ordinal" `of the first test case starts at 0, and the rest of the test cases are off by one.
**Environment**:
Windows 11/Ubuntu 20.04
**Log Info**:
**Screenshots**: if applicable, please add attachments showing screenshots to help explain the issue.
**Additional context**:
```
{"type":"runs", "id":"pc2-4141", "op":"create", "data": {"id":"TestCase-2971170821107431354","judgement_id":"Run-25513900955186117","ordinal":0,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.607-05","contest_time":"00:25:52.371"}}
{"type":"runs", "id":"pc2-4142", "op":"create", "data": {"id":"TestCase--1950512010137499320","judgement_id":"Run-25513900955186117","ordinal":1,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.656-05","contest_time":"00:25:52.420"}}
{"type":"runs", "id":"pc2-4143", "op":"create", "data": {"id":"TestCase--7824024682353058102","judgement_id":"Run-25513900955186117","ordinal":2,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.726-05","contest_time":"00:25:52.490"}}
{"type":"runs", "id":"pc2-4144", "op":"create", "data": {"id":"TestCase-5245915230241241075","judgement_id":"Run-25513900955186117","ordinal":3,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.798-05","contest_time":"00:25:52.562"}}
```
|
1.0
|
event-feed API endpoint returns incorrect ordinal field for "runs" testcases -
**Describe the issue**:
Reported by DOMJudge: For `"runs" `test cases in the event-feed, the `"ordinal"` value should start at 1 not 0 as per the CLICS Specification. Currently, the first test case has an ordinal value of `"0"`.
**To Reproduce**:
Start up a contest that has some judged runs in it.
Start the Event Feed.
Connect to the event-feed endpoint and download the event feed.
Note that the first testcase for each submission has an `"ordinal"` value of 0. Consequently, all following test cases are off by one for `"ordinal"`
**Expected behavior**:
The `"ordinal" `value of the first test case should start at 1, not 0.
**Actual behavior**:
The `"ordinal" `of the first test case starts at 0, and the rest of the test cases are off by one.
**Environment**:
Windows 11/Ubuntu 20.04
**Log Info**:
**Screenshots**: if applicable, please add attachments showing screenshots to help explain the issue.
**Additional context**:
```
{"type":"runs", "id":"pc2-4141", "op":"create", "data": {"id":"TestCase-2971170821107431354","judgement_id":"Run-25513900955186117","ordinal":0,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.607-05","contest_time":"00:25:52.371"}}
{"type":"runs", "id":"pc2-4142", "op":"create", "data": {"id":"TestCase--1950512010137499320","judgement_id":"Run-25513900955186117","ordinal":1,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.656-05","contest_time":"00:25:52.420"}}
{"type":"runs", "id":"pc2-4143", "op":"create", "data": {"id":"TestCase--7824024682353058102","judgement_id":"Run-25513900955186117","ordinal":2,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.726-05","contest_time":"00:25:52.490"}}
{"type":"runs", "id":"pc2-4144", "op":"create", "data": {"id":"TestCase-5245915230241241075","judgement_id":"Run-25513900955186117","ordinal":3,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.798-05","contest_time":"00:25:52.562"}}
```
|
non_process
|
event feed api endpoint returns incorrect ordinal field for runs testcases describe the issue reported by domjudge for runs test cases in the event feed the ordinal value should start at not as per the clics specification currently the first test case has an ordinal value of to reproduce start up a contest that has some judged runs in it start the event feed connect to the event feed endpoint and download the event feed note that the first testcase for each submission has an ordinal value of consequently all following test cases are off by one for ordinal expected behavior the ordinal value of the first test case should start at not actual behavior the ordinal of the first test case starts at and the rest of the test cases are off by one environment windows ubuntu log info screenshots if applicable please add attachments showing screenshots to help explain the issue additional context type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time
| 0
|
4,802
| 7,696,239,052
|
IssuesEvent
|
2018-05-18 14:44:54
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Refactor 'Code instrumentation' subsystem
|
!IMPORTANT! AREA: client AREA: server SYSTEM: resource processing TYPE: enhancement
|
We will use a native getter from property descriptor.
It allows to drop server and client script processing, improves stack traces for client errors and etc.
It will be done after updating mobile browsers `Android 6.0+` and `Safari 10+`.
|
1.0
|
Refactor 'Code instrumentation' subsystem - We will use a native getter from property descriptor.
It allows to drop server and client script processing, improves stack traces for client errors and etc.
It will be done after updating mobile browsers `Android 6.0+` and `Safari 10+`.
|
process
|
refactor code instrumentation subsystem we will use a native getter from property descriptor it allows to drop server and client script processing improves stack traces for client errors and etc it will be done after updating mobile browsers android and safari
| 1
|
137,035
| 12,743,816,124
|
IssuesEvent
|
2020-06-26 11:12:55
|
dask/dask-jobqueue
|
https://api.github.com/repos/dask/dask-jobqueue
|
closed
|
New Github design shows oldest release as latest?
|
documentation
|
The new side bar seems to get the order of releases wrong. It lists the v0.1 as the latest release:

(Not sure if this is something that can be fixed in this repo or if this is a bug with Github.)
|
1.0
|
New Github design shows oldest release as latest? - The new side bar seems to get the order of releases wrong. It lists the v0.1 as the latest release:

(Not sure if this is something that can be fixed in this repo or if this is a bug with Github.)
|
non_process
|
new github design shows oldest release as latest the new side bar seems to get the order of releases wrong it lists the as the latest release not sure if this is something that can be fixed in this repo or if this is a bug with github
| 0
|
69,694
| 17,792,168,391
|
IssuesEvent
|
2021-08-31 17:29:18
|
netlify/cli
|
https://api.github.com/repos/netlify/cli
|
closed
|
Support ODB in local builds and `netlify dev`
|
type: feature proj/odb-build
|
ODB should work in `netlify build` (local builds) and in `netlify dev`.
This is a cover issue. More details should come as we implement the production behavior.
See also: https://github.com/netlify/buildbot/issues/1645, https://github.com/netlify/buildbot/issues/1646, https://github.com/netlify/build/issues/3547, https://github.com/netlify/build/issues/3548, https://github.com/netlify/build/issues/3549, https://github.com/netlify/build/issues/3550
|
1.0
|
Support ODB in local builds and `netlify dev` - ODB should work in `netlify build` (local builds) and in `netlify dev`.
This is a cover issue. More details should come as we implement the production behavior.
See also: https://github.com/netlify/buildbot/issues/1645, https://github.com/netlify/buildbot/issues/1646, https://github.com/netlify/build/issues/3547, https://github.com/netlify/build/issues/3548, https://github.com/netlify/build/issues/3549, https://github.com/netlify/build/issues/3550
|
non_process
|
support odb in local builds and netlify dev odb should work in netlify build local builds and in netlify dev this is a cover issue more details should come as we implement the production behavior see also
| 0
|
286,864
| 31,773,270,358
|
IssuesEvent
|
2023-09-12 13:04:28
|
ssasy-auth/core
|
https://api.github.com/repos/ssasy-auth/core
|
closed
|
enhance challenge solution
|
question security
|
To solve a challenge, the current solution is to hash the challenge's nonce. An improvement would be to:
- [ ] hash the entire challenge (without the solution property).
|
True
|
enhance challenge solution - To solve a challenge, the current solution is to hash the challenge's nonce. An improvement would be to:
- [ ] hash the entire challenge (without the solution property).
|
non_process
|
enhance challenge solution to solve a challenge the current solution is to hash the challenge s nonce an improvement would be to hash the entire challenge without the solution property
| 0
|
153,366
| 24,118,325,343
|
IssuesEvent
|
2022-09-20 16:23:35
|
carbon-design-system/carbon-platform
|
https://api.github.com/repos/carbon-design-system/carbon-platform
|
opened
|
Resource card usage — When to use regular size versus small size?
|
role: design 🎨
|
🎨 Visual
**Page:** https://next.carbondesignsystem.com/designing/get-started
**Issue:** Resource card usage — When to use regular size versus small size?
**Suggestion:** I was wondering if there are currently any rules around when to use each size of the Resource card component. On the `Designing > Get started` page, Step 2 (Access design tools) and Step 3 (tutorials) use different size resource cards. Was not sure if this was intentional or not.
Looks like that maybe we are using the regular size resource card for the most important cards on the page (Design tools) and we use small ones for secondary link out content (Tutorials). Should they or shouldn't they coexist on the same page? I think we just need to outline some guidelines one when we use them and how we use them.
----
**Step 2 (Access design tools)**
<img width="818" alt="Screen Shot 2022-09-19 at 4 17 31 PM" src="https://user-images.githubusercontent.com/43969356/191108663-05d4d15f-3799-4c49-8347-0ec8c037fb46.png">
**Step 3 (tutorials)**
<img width="781" alt="Screen Shot 2022-09-19 at 4 17 42 PM" src="https://user-images.githubusercontent.com/43969356/191108688-6e3dc60f-2333-4fed-85a0-31ddd668ad45.png">
|
1.0
|
Resource card usage — When to use regular size versus small size? - 🎨 Visual
**Page:** https://next.carbondesignsystem.com/designing/get-started
**Issue:** Resource card usage — When to use regular size versus small size?
**Suggestion:** I was wondering if there are currently any rules around when to use each size of the Resource card component. On the `Designing > Get started` page, Step 2 (Access design tools) and Step 3 (tutorials) use different size resource cards. Was not sure if this was intentional or not.
Looks like that maybe we are using the regular size resource card for the most important cards on the page (Design tools) and we use small ones for secondary link out content (Tutorials). Should they or shouldn't they coexist on the same page? I think we just need to outline some guidelines one when we use them and how we use them.
----
**Step 2 (Access design tools)**
<img width="818" alt="Screen Shot 2022-09-19 at 4 17 31 PM" src="https://user-images.githubusercontent.com/43969356/191108663-05d4d15f-3799-4c49-8347-0ec8c037fb46.png">
**Step 3 (tutorials)**
<img width="781" alt="Screen Shot 2022-09-19 at 4 17 42 PM" src="https://user-images.githubusercontent.com/43969356/191108688-6e3dc60f-2333-4fed-85a0-31ddd668ad45.png">
|
non_process
|
resource card usage — when to use regular size versus small size 🎨 visual page issue resource card usage — when to use regular size versus small size suggestion i was wondering if there are currently any rules around when to use each size of the resource card component on the designing get started page step access design tools and step tutorials use different size resource cards was not sure if this was intentional or not looks like that maybe we are using the regular size resource card for the most important cards on the page design tools and we use small ones for secondary link out content tutorials should they or shouldn t they coexist on the same page i think we just need to outline some guidelines one when we use them and how we use them step access design tools img width alt screen shot at pm src step tutorials img width alt screen shot at pm src
| 0
|
268,858
| 8,415,266,657
|
IssuesEvent
|
2018-10-13 12:55:45
|
ngageoint/hootenanny
|
https://api.github.com/repos/ngageoint/hootenanny
|
closed
|
Symmetric translation fixes
|
Category: Translation Priority: High Status: In Progress Type: Bug Type: Support in progress
|
Initial list to look at:
* Motor vehicle station
* Shopping complex
|
1.0
|
Symmetric translation fixes - Initial list to look at:
* Motor vehicle station
* Shopping complex
|
non_process
|
symmetric translation fixes initial list to look at motor vehicle station shopping complex
| 0
|
16,514
| 2,909,794,534
|
IssuesEvent
|
2015-06-21 03:30:24
|
prettydiff/prettydiff
|
https://api.github.com/repos/prettydiff/prettydiff
|
closed
|
XML comparison sometimes gives erroneous errors with attributes in different order
|
Defect QA Underway
|
Prettydiff usually does a good job of ignoring differences just in the order of attributes. However, I've found some cases where it doesn't work. In the output, the attributes do not appear to be sorted on the offending lines.
Here is an example gist containing two files to diff:
https://gist.github.com/murrayju/5d99c370e06176044fa0
(I'm confident that they are equal, oXygen's diff tool agrees)
|
1.0
|
XML comparison sometimes gives erroneous errors with attributes in different order - Prettydiff usually does a good job of ignoring differences just in the order of attributes. However, I've found some cases where it doesn't work. In the output, the attributes do not appear to be sorted on the offending lines.
Here is an example gist containing two files to diff:
https://gist.github.com/murrayju/5d99c370e06176044fa0
(I'm confident that they are equal, oXygen's diff tool agrees)
|
non_process
|
xml comparison sometimes gives erroneous errors with attributes in different order prettydiff usually does a good job of ignoring differences just in the order of attributes however i ve found some cases where it doesn t work in the output the attributes do not appear to be sorted on the offending lines here is an example gist containing two files to diff i m confident that they are equal oxygen s diff tool agrees
| 0
|
181,923
| 21,664,468,081
|
IssuesEvent
|
2022-05-07 01:27:02
|
eldorplus/portfolio
|
https://api.github.com/repos/eldorplus/portfolio
|
closed
|
CVE-2018-11695 (High) detected in opennms-opennms-source-23.0.3-1 - autoclosed
|
security vulnerability
|
## CVE-2018-11695 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-23.0.3-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/eldorplus/portfolio/commits/d39c6030d1112cc864fbcffa04099b507e753f36">d39c6030d1112cc864fbcffa04099b507e753f36</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (13)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /portfolio/node_modules/node-sass/src/sass_context_wrapper.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/expand.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/parser.hpp
- /portfolio/node_modules/node-sass/src/libsass/src/util.hpp
- /portfolio/node_modules/node-sass/src/libsass/src/cssize.cpp
- /portfolio/node_modules/node-gyp/gyp/pylib/gyp/MSVSUtil.py
- /portfolio/node_modules/node-sass/src/libsass/src/functions.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /portfolio/node_modules/node-sass/src/callback_bridge.h
- /portfolio/node_modules/node-sass/src/libsass/src/sass.cpp
- /portfolio/node_modules/node-sass/src/sass_context_wrapper.h
- /portfolio/node_modules/node-sass/src/libsass/src/eval.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/debugger.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.2. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695>CVE-2018-11695</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-11695 (High) detected in opennms-opennms-source-23.0.3-1 - autoclosed - ## CVE-2018-11695 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-23.0.3-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/eldorplus/portfolio/commits/d39c6030d1112cc864fbcffa04099b507e753f36">d39c6030d1112cc864fbcffa04099b507e753f36</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (13)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /portfolio/node_modules/node-sass/src/sass_context_wrapper.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/expand.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/parser.hpp
- /portfolio/node_modules/node-sass/src/libsass/src/util.hpp
- /portfolio/node_modules/node-sass/src/libsass/src/cssize.cpp
- /portfolio/node_modules/node-gyp/gyp/pylib/gyp/MSVSUtil.py
- /portfolio/node_modules/node-sass/src/libsass/src/functions.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /portfolio/node_modules/node-sass/src/callback_bridge.h
- /portfolio/node_modules/node-sass/src/libsass/src/sass.cpp
- /portfolio/node_modules/node-sass/src/sass_context_wrapper.h
- /portfolio/node_modules/node-sass/src/libsass/src/eval.cpp
- /portfolio/node_modules/node-sass/src/libsass/src/debugger.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.2. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695>CVE-2018-11695</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in opennms opennms source autoclosed cve high severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries portfolio node modules node sass src sass context wrapper cpp portfolio node modules node sass src libsass src expand cpp portfolio node modules node sass src libsass src parser hpp portfolio node modules node sass src libsass src util hpp portfolio node modules node sass src libsass src cssize cpp portfolio node modules node gyp gyp pylib gyp msvsutil py portfolio node modules node sass src libsass src functions cpp portfolio node modules node sass src libsass src prelexer cpp portfolio node modules node sass src callback bridge h portfolio node modules node sass src libsass src sass cpp portfolio node modules node sass src sass context wrapper h portfolio node modules node sass src libsass src eval cpp portfolio node modules node sass src libsass src debugger hpp vulnerability details an issue was discovered in libsass through a null pointer dereference was found in the function sass expand operator which could be leveraged by an attacker to cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
| 0
|
804,267
| 29,481,765,826
|
IssuesEvent
|
2023-06-02 06:29:22
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.microsoft.com - design is broken
|
browser-chrome priority-critical
|
<!-- @browser: Chrome 114.0.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36 -->
<!-- @reported_with: unknown -->
**URL**: https://www.microsoft.com/en-us/edge?form=MA13FJ&exp=e00
**Browser / Version**: Chrome 114.0.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Design is broken
**Description**: Items are misaligned
**Steps to Reproduce**:
Design is broken on the navbar.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/c99633d1-b0de-469e-9b32-3cb72764a337.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.microsoft.com - design is broken - <!-- @browser: Chrome 114.0.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36 -->
<!-- @reported_with: unknown -->
**URL**: https://www.microsoft.com/en-us/edge?form=MA13FJ&exp=e00
**Browser / Version**: Chrome 114.0.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Design is broken
**Description**: Items are misaligned
**Steps to Reproduce**:
Design is broken on the navbar.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/c99633d1-b0de-469e-9b32-3cb72764a337.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
design is broken url browser version chrome operating system windows tested another browser yes edge problem type design is broken description items are misaligned steps to reproduce design is broken on the navbar view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
16,592
| 21,641,748,591
|
IssuesEvent
|
2022-05-05 19:34:06
|
MagicalAlexey/Keepers-of-the-Stones
|
https://api.github.com/repos/MagicalAlexey/Keepers-of-the-Stones
|
opened
|
Adding a full-fledged level system
|
new feature In process
|
There are plans to add a full-fledged pumping system of stones in order to improve the combat performance of stones.
Tasks:
- [ ] Add GUI skills and levels
- [ ] Add an Element Coin
- [ ] Add a way to get a coin
- [ ] Add additional abilities
|
1.0
|
Adding a full-fledged level system - There are plans to add a full-fledged pumping system of stones in order to improve the combat performance of stones.
Tasks:
- [ ] Add GUI skills and levels
- [ ] Add an Element Coin
- [ ] Add a way to get a coin
- [ ] Add additional abilities
|
process
|
adding a full fledged level system there are plans to add a full fledged pumping system of stones in order to improve the combat performance of stones tasks add gui skills and levels add an element coin add a way to get a coin add additional abilities
| 1
|
404,828
| 11,863,498,011
|
IssuesEvent
|
2020-03-25 19:50:20
|
near/near-api-js
|
https://api.github.com/repos/near/near-api-js
|
opened
|
Expose `genesis_config` and `genesis_records` RPC API
|
Priority 3 enhancement good first issue
|
We have recently added (nearprotocol/nearcore#2109) new RPC API methods: genesis_config and genesis_records (with `EXPERIMENTAL_` prefix).
The API documentation is currently defined in the PR description mentioned above.
|
1.0
|
Expose `genesis_config` and `genesis_records` RPC API - We have recently added (nearprotocol/nearcore#2109) new RPC API methods: genesis_config and genesis_records (with `EXPERIMENTAL_` prefix).
The API documentation is currently defined in the PR description mentioned above.
|
non_process
|
expose genesis config and genesis records rpc api we have recently added nearprotocol nearcore new rpc api methods genesis config and genesis records with experimental prefix the api documentation is currently defined in the pr description mentioned above
| 0
|
53,886
| 6,772,631,440
|
IssuesEvent
|
2017-10-27 00:07:19
|
HabitRPG/habitica
|
https://api.github.com/repos/HabitRPG/habitica
|
opened
|
Glasses and animal ears are not working
|
hacktoberfest help wanted POST-REDESIGN priority: important sections: Avatar/User Modal
|
Glasses and animal ears are not working. This may be happening because the Edit Avatar modal is equipping them to Battle Gear even when the user has Costume enabled.
|
1.0
|
Glasses and animal ears are not working - Glasses and animal ears are not working. This may be happening because the Edit Avatar modal is equipping them to Battle Gear even when the user has Costume enabled.
|
non_process
|
glasses and animal ears are not working glasses and animal ears are not working this may be happening because the edit avatar modal is equipping them to battle gear even when the user has costume enabled
| 0
|
21,096
| 28,045,730,679
|
IssuesEvent
|
2023-03-28 22:39:20
|
apache/arrow-rs
|
https://api.github.com/repos/apache/arrow-rs
|
closed
|
Make ObjectMeta Non Exhaustive
|
enhancement development-process
|
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.**
<!--
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
(This section helps Arrow developers understand the context and *why* for this feature, in addition to the *what*)
-->
Currently the fields on `ObjectMeta` are public, making a breaking change to add additional metadata, such as in https://github.com/apache/arrow-rs/issues/2240
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen.
-->
I would like to make `ObjectMeta` [non_exhaustive](https://doc.rust-lang.org/reference/attributes/type_system.html#the-non_exhaustive-attribute) to allow adding new fields without requiring a breaking change.
This will require adding a constructor to `ObjectMeta` to allow it to be safely constructed outside of the crate.
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
|
1.0
|
Make ObjectMeta Non Exhaustive - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.**
<!--
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
(This section helps Arrow developers understand the context and *why* for this feature, in addition to the *what*)
-->
Currently the fields on `ObjectMeta` are public, making a breaking change to add additional metadata, such as in https://github.com/apache/arrow-rs/issues/2240
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen.
-->
I would like to make `ObjectMeta` [non_exhaustive](https://doc.rust-lang.org/reference/attributes/type_system.html#the-non_exhaustive-attribute) to allow adding new fields without requiring a breaking change.
This will require adding a constructor to `ObjectMeta` to allow it to be safely constructed outside of the crate.
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
|
process
|
make objectmeta non exhaustive is your feature request related to a problem or challenge please describe what you are trying to do a clear and concise description of what the problem is ex i m always frustrated when this section helps arrow developers understand the context and why for this feature in addition to the what currently the fields on objectmeta are public making a breaking change to add additional metadata such as in describe the solution you d like a clear and concise description of what you want to happen i would like to make objectmeta to allow adding new fields without requiring a breaking change this will require adding a constructor to objectmeta to allow it to be safely constructed outside of the crate describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
| 1
|
14,521
| 17,619,035,590
|
IssuesEvent
|
2021-08-18 13:22:00
|
linked-art/linked.art
|
https://api.github.com/repos/linked-art/linked.art
|
closed
|
Should we version the spec?
|
model defer process
|
I think semantic versioning of this would be really useful at some point.
(as a 0.1 version, so we can make breaking changes as needed, but it would be helpful to indicate that this is in flux to people, and also to indicate when it _stops_ being in flux)
|
1.0
|
Should we version the spec? - I think semantic versioning of this would be really useful at some point.
(as a 0.1 version, so we can make breaking changes as needed, but it would be helpful to indicate that this is in flux to people, and also to indicate when it _stops_ being in flux)
|
process
|
should we version the spec i think semantic versioning of this would be really useful at some point as a version so we can make breaking changes as needed but it would be helpful to indicate that this is in flux to people and also to indicate when it stops being in flux
| 1
|
37,882
| 15,391,048,812
|
IssuesEvent
|
2021-03-03 14:08:39
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[App Arch] Can't set Lucene Query String options through Advanced Settings
|
:AppServices/fix-it-week Feature:Query Bar Feature:Search Team:AppServices bug impact:high loe:hours triaged
|
If you try to add any query string options other than `analyze_wildcard` through Advanced Settings, you will get an error like
```
[validation [query:queryString:options].lenient]: definition for this key is missing
```
It appears that the schema for that setting only has the `analyze_wildcard` as a valid key and not any of the other valid query string options.
https://github.com/elastic/kibana/blob/master/src/plugins/data/server/ui_settings.ts#L106
|
2.0
|
[App Arch] Can't set Lucene Query String options through Advanced Settings - If you try to add any query string options other than `analyze_wildcard` through Advanced Settings, you will get an error like
```
[validation [query:queryString:options].lenient]: definition for this key is missing
```
It appears that the schema for that setting only has the `analyze_wildcard` as a valid key and not any of the other valid query string options.
https://github.com/elastic/kibana/blob/master/src/plugins/data/server/ui_settings.ts#L106
|
non_process
|
can t set lucene query string options through advanced settings if you try to add any query string options other than analyze wildcard through advanced settings you will get an error like lenient definition for this key is missing it appears that the schema for that setting only has the analyze wildcard as a valid key and not any of the other valid query string options
| 0
|
188,322
| 22,046,298,749
|
IssuesEvent
|
2022-05-30 02:22:15
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
closed
|
CVE-2016-10906 (High) detected in fedorav4.6 - autoclosed
|
security vulnerability
|
## CVE-2016-10906 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fedorav4.6</b></p></summary>
<p>
<p>Fedora kernel git tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git>https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in drivers/net/ethernet/arc/emac_main.c in the Linux kernel before 4.5. A use-after-free is caused by a race condition between the functions arc_emac_tx and arc_emac_tx_clean.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10906>CVE-2016-10906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906</a></p>
<p>Release Date: 2019-08-19</p>
<p>Fix Resolution: v4.5-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-10906 (High) detected in fedorav4.6 - autoclosed - ## CVE-2016-10906 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fedorav4.6</b></p></summary>
<p>
<p>Fedora kernel git tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git>https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in drivers/net/ethernet/arc/emac_main.c in the Linux kernel before 4.5. A use-after-free is caused by a race condition between the functions arc_emac_tx and arc_emac_tx_clean.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10906>CVE-2016-10906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906</a></p>
<p>Release Date: 2019-08-19</p>
<p>Fix Resolution: v4.5-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in autoclosed cve high severity vulnerability vulnerable library fedora kernel git tree library home page a href found in head commit a href found in base branch master vulnerable source files drivers net ethernet arc emac main c drivers net ethernet arc emac main c vulnerability details an issue was discovered in drivers net ethernet arc emac main c in the linux kernel before a use after free is caused by a race condition between the functions arc emac tx and arc emac tx clean publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
101,639
| 4,127,679,637
|
IssuesEvent
|
2016-06-10 00:11:38
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
opened
|
Refactor SELinux related volume code in kubelet
|
area/storage priority/P2 team/cluster
|
Refactor SELinux related volume code in kubelet. For example the the code depends on being able to mutate the pod object.
|
1.0
|
Refactor SELinux related volume code in kubelet - Refactor SELinux related volume code in kubelet. For example the the code depends on being able to mutate the pod object.
|
non_process
|
refactor selinux related volume code in kubelet refactor selinux related volume code in kubelet for example the the code depends on being able to mutate the pod object
| 0
|
42,700
| 22,780,339,686
|
IssuesEvent
|
2022-07-08 18:57:10
|
appsmithorg/appsmith
|
https://api.github.com/repos/appsmithorg/appsmith
|
opened
|
[Perf]: Move Form widget's Child widgets logic to derived properties
|
Performance Form Widget Task List Widget UI Performance
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### SubTasks
Form Widget uses childWidgets to get values to check each child's value, isValid, isDirty on the main Thread. This should Ideally be done in worker thread through derived Properties. This also helps us avoid Form widget from needing childWidgets which might also cause it to Re-render.
This can also be tried for List Widget.
|
True
|
[Perf]: Move Form widget's Child widgets logic to derived properties - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### SubTasks
Form Widget uses childWidgets to get values to check each child's value, isValid, isDirty on the main Thread. This should Ideally be done in worker thread through derived Properties. This also helps us avoid Form widget from needing childWidgets which might also cause it to Re-render.
This can also be tried for List Widget.
|
non_process
|
move form widget s child widgets logic to derived properties is there an existing issue for this i have searched the existing issues subtasks form widget uses childwidgets to get values to check each child s value isvalid isdirty on the main thread this should ideally be done in worker thread through derived properties this also helps us avoid form widget from needing childwidgets which might also cause it to re render this can also be tried for list widget
| 0
|
17,459
| 23,282,747,984
|
IssuesEvent
|
2022-08-05 13:39:38
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag')
|
bug/1-unconfirmed kind/bug process/candidate tech/typescript topic: transaction team/client
|
### Bug description
I try use prisma transaction and have error:
```
<project-home>/src/database/client/runtime/index.js:23284
const hasPreview = engine._hasPreviewFlag("tracing");
^
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag')
at getTracingConfig (<project-home>/src/database/client/runtime/index.js:23284:29)
at Object.$transaction [as transaction] (<project-home>/src/database/client/runtime/index.js:29563:29)
```
### How to reproduce
Simple code in my handler:
```js
await prisma.$transaction([
prisma.passwordOfAdministrator.upsert({
where: {
administratorId: confirmation.administratorId,
},
create: {
administratorId: confirmation.administratorId,
hash: password.hash,
salt: password.salt,
createdAt,
updatedAt,
},
update: {
hash: password.hash,
salt: password.salt,
updatedAt,
},
}),
prisma.recoveryPasswordConfirmation.update({
where: {
id: confirmation.id,
},
data: {
ip,
agent,
confirmAt,
},
}),
])
```
### Expected behavior
Execute `prisma.$transaction` without error:
```
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag')
```
### Prisma information
```
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
output = "./client"
engineType = "binary"
}
```
or
```
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
output = "./client"
engineType = "library"
}
```
### Environment & setup
- OS: [Mac OS]
- Database: [PostgreSQL]
- Node.js version: [v16.15.0]
### Prisma Version
```
prisma : 4.1.1
@prisma/client : 4.1.1
Current platform : darwin
Query Engine (Node-API) : libquery-engine 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node)
Migration Engine : migration-engine-cli 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 8d8414deb360336e4698a65aa45a1fbaf1ce13d8
Studio : 0.469.0
```
|
1.0
|
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag') - ### Bug description
I try use prisma transaction and have error:
```
<project-home>/src/database/client/runtime/index.js:23284
const hasPreview = engine._hasPreviewFlag("tracing");
^
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag')
at getTracingConfig (<project-home>/src/database/client/runtime/index.js:23284:29)
at Object.$transaction [as transaction] (<project-home>/src/database/client/runtime/index.js:29563:29)
```
### How to reproduce
Simple code in my handler:
```js
await prisma.$transaction([
prisma.passwordOfAdministrator.upsert({
where: {
administratorId: confirmation.administratorId,
},
create: {
administratorId: confirmation.administratorId,
hash: password.hash,
salt: password.salt,
createdAt,
updatedAt,
},
update: {
hash: password.hash,
salt: password.salt,
updatedAt,
},
}),
prisma.recoveryPasswordConfirmation.update({
where: {
id: confirmation.id,
},
data: {
ip,
agent,
confirmAt,
},
}),
])
```
### Expected behavior
Execute `prisma.$transaction` without error:
```
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag')
```
### Prisma information
```
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
output = "./client"
engineType = "binary"
}
```
or
```
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
output = "./client"
engineType = "library"
}
```
### Environment & setup
- OS: [Mac OS]
- Database: [PostgreSQL]
- Node.js version: [v16.15.0]
### Prisma Version
```
prisma : 4.1.1
@prisma/client : 4.1.1
Current platform : darwin
Query Engine (Node-API) : libquery-engine 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node)
Migration Engine : migration-engine-cli 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 8d8414deb360336e4698a65aa45a1fbaf1ce13d8
Studio : 0.469.0
```
|
process
|
typeerror cannot read properties of undefined reading haspreviewflag bug description i try use prisma transaction and have error src database client runtime index js const haspreview engine haspreviewflag tracing typeerror cannot read properties of undefined reading haspreviewflag at gettracingconfig src database client runtime index js at object transaction src database client runtime index js how to reproduce simple code in my handler js await prisma transaction prisma passwordofadministrator upsert where administratorid confirmation administratorid create administratorid confirmation administratorid hash password hash salt password salt createdat updatedat update hash password hash salt password salt updatedat prisma recoverypasswordconfirmation update where id confirmation id data ip agent confirmat expected behavior execute prisma transaction without error typeerror cannot read properties of undefined reading haspreviewflag prisma information datasource db provider postgresql url env database url generator client provider prisma client js output client enginetype binary or datasource db provider postgresql url env database url generator client provider prisma client js output client enginetype library environment setup os database node js version prisma version prisma prisma client current platform darwin query engine node api libquery engine at node modules prisma engines libquery engine darwin dylib node migration engine migration engine cli at node modules prisma engines migration engine darwin introspection engine introspection core at node modules prisma engines introspection engine darwin format binary prisma fmt at node modules prisma engines prisma fmt darwin default engines hash studio
| 1
|
435,263
| 12,533,480,357
|
IssuesEvent
|
2020-06-04 17:40:43
|
inverse-inc/packetfence
|
https://api.github.com/repos/inverse-inc/packetfence
|
closed
|
Auth source rules are disabled by default
|
Priority: High Type: Bug
|
**Describe the bug**
The auth source rules are shown as disabled by default. This should be the opposite
**To Reproduce**
1. Create an AD source with an auth rule
2. The auth rule shows disabled by default
|
1.0
|
Auth source rules are disabled by default - **Describe the bug**
The auth source rules are shown as disabled by default. This should be the opposite
**To Reproduce**
1. Create an AD source with an auth rule
2. The auth rule shows disabled by default
|
non_process
|
auth source rules are disabled by default describe the bug the auth source rules are shown as disabled by default this should be the opposite to reproduce create an ad source with an auth rule the auth rule shows disabled by default
| 0
|
12,808
| 3,653,103,899
|
IssuesEvent
|
2016-02-17 05:01:59
|
Esri/angular-esri-map
|
https://api.github.com/repos/Esri/angular-esri-map
|
closed
|
Update scene view example pages to show warning for unsupported (mobile) devices
|
documentation enhancement
|
depends on #239
See this sample page for the pattern:
https://developers.arcgis.com/javascript/beta/sample-code/scene-webgl-support/index.html
|
1.0
|
Update scene view example pages to show warning for unsupported (mobile) devices - depends on #239
See this sample page for the pattern:
https://developers.arcgis.com/javascript/beta/sample-code/scene-webgl-support/index.html
|
non_process
|
update scene view example pages to show warning for unsupported mobile devices depends on see this sample page for the pattern
| 0
|
101,745
| 11,255,951,472
|
IssuesEvent
|
2020-01-12 13:08:33
|
ManoukK/NIBUD_datavisualisatie
|
https://api.github.com/repos/ManoukK/NIBUD_datavisualisatie
|
opened
|
README maken
|
documentation
|
readme make en afmaken
Hierin moet ook staan hoe het werkt met de dataset van het HIBUD omdat die niet online kan
|
1.0
|
README maken - readme make en afmaken
Hierin moet ook staan hoe het werkt met de dataset van het HIBUD omdat die niet online kan
|
non_process
|
readme maken readme make en afmaken hierin moet ook staan hoe het werkt met de dataset van het hibud omdat die niet online kan
| 0
|
186,939
| 15,088,052,655
|
IssuesEvent
|
2021-02-05 23:36:43
|
dapr/php-sdk
|
https://api.github.com/repos/dapr/php-sdk
|
closed
|
Add an editorconfig
|
documentation enhancement
|
Formatting is kind-of spotty and inconsistent. It'd be good to have some consistency and have formatting be enforced in PRs.
|
1.0
|
Add an editorconfig - Formatting is kind-of spotty and inconsistent. It'd be good to have some consistency and have formatting be enforced in PRs.
|
non_process
|
add an editorconfig formatting is kind of spotty and inconsistent it d be good to have some consistency and have formatting be enforced in prs
| 0
|
20,364
| 27,021,106,449
|
IssuesEvent
|
2023-02-11 02:25:13
|
openxla/stablehlo
|
https://api.github.com/repos/openxla/stablehlo
|
opened
|
Populate conformance suite with "interesting" test cases
|
Process
|
Our current methodology for testing the reference implementation involves writing a minimal amount of tests (https://github.com/openxla/stablehlo/blob/main/docs/reference_checklist.md), but for a conformance suite we'll likely want to write "interesting" tests that cover potentially surprising behavior, even if strictly speaking these tests don't increase coverage of the codebase. Let's think about how this could look like.
|
1.0
|
Populate conformance suite with "interesting" test cases - Our current methodology for testing the reference implementation involves writing a minimal amount of tests (https://github.com/openxla/stablehlo/blob/main/docs/reference_checklist.md), but for a conformance suite we'll likely want to write "interesting" tests that cover potentially surprising behavior, even if strictly speaking these tests don't increase coverage of the codebase. Let's think about how this could look like.
|
process
|
populate conformance suite with interesting test cases our current methodology for testing the reference implementation involves writing a minimal amount of tests but for a conformance suite we ll likely want to write interesting tests that cover potentially surprising behavior even if strictly speaking these tests don t increase coverage of the codebase let s think about how this could look like
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.