Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
167,036
14,099,697,038
IssuesEvent
2020-11-06 02:06:03
matplotlib/matplotlib
https://api.github.com/repos/matplotlib/matplotlib
opened
Current `tight_layout` example raises UserWarning
API Changes Documentation
<!--To help us understand and resolve your issue, please fill out the form to the best of your ability.--> <!--You can feel free to delete the sections that do not apply.--> ### Problem As can be seen in the [tight layout docs](https://matplotlib.org/tutorials/intermediate/tight_layout_guide.html) (Ctrl-F for "Out:"), `GridSpec.tight_layout` raises a `UserWarning` whenever there are `Axes` in a figure that come from other `GridSpec` besides `self`. Is this the intended API? If so, is the use case being demonstrated when that warning is raised (using different subplot specs to render in different `rect`'s within a single figure, without needed nested gridspec's) deprecated? Should we be instead encouraging users to use nested GridSpec's instead for this use-case? If not, then should there be a flag to `tight_layout` that tells it to ignore other `Axes` in the figure, like a "I know what I'm doing" flag? I can't see a way for `tight_layout.get_subplotspec_list` to differentiate, currently, (and this may be due to a lack of my own imagination) between the two relevant options: 1. the user accidentally passed in a subplotspec that doesn't encompass the entire figure (I guess that's what the warning is for), OR 2. whether they intentionally are passing in gridspec that only describes a subset of Axes (as it's being used in the example)
1.0
Current `tight_layout` example raises UserWarning - <!--To help us understand and resolve your issue, please fill out the form to the best of your ability.--> <!--You can feel free to delete the sections that do not apply.--> ### Problem As can be seen in the [tight layout docs](https://matplotlib.org/tutorials/intermediate/tight_layout_guide.html) (Ctrl-F for "Out:"), `GridSpec.tight_layout` raises a `UserWarning` whenever there are `Axes` in a figure that come from other `GridSpec` besides `self`. Is this the intended API? If so, is the use case being demonstrated when that warning is raised (using different subplot specs to render in different `rect`'s within a single figure, without needed nested gridspec's) deprecated? Should we be instead encouraging users to use nested GridSpec's instead for this use-case? If not, then should there be a flag to `tight_layout` that tells it to ignore other `Axes` in the figure, like a "I know what I'm doing" flag? I can't see a way for `tight_layout.get_subplotspec_list` to differentiate, currently, (and this may be due to a lack of my own imagination) between the two relevant options: 1. the user accidentally passed in a subplotspec that doesn't encompass the entire figure (I guess that's what the warning is for), OR 2. whether they intentionally are passing in gridspec that only describes a subset of Axes (as it's being used in the example)
non_process
current tight layout example raises userwarning problem as can be seen in the ctrl f for out gridspec tight layout raises a userwarning whenever there are axes in a figure that come from other gridspec besides self is this the intended api if so is the use case being demonstrated when that warning is raised using different subplot specs to render in different rect s within a single figure without needed nested gridspec s deprecated should we be instead encouraging users to use nested gridspec s instead for this use case if not then should there be a flag to tight layout that tells it to ignore other axes in the figure like a i know what i m doing flag i can t see a way for tight layout get subplotspec list to differentiate currently and this may be due to a lack of my own imagination between the two relevant options the user accidentally passed in a subplotspec that doesn t encompass the entire figure i guess that s what the warning is for or whether they intentionally are passing in gridspec that only describes a subset of axes as it s being used in the example
0
39,926
10,421,705,655
IssuesEvent
2019-09-16 07:06:03
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Fix Javadoc warnings
C: Build C: Documentation E: All Editions P: Low T: Defect
The build currently complains about the following Javadoc warnings: ``` [WARNING] Javadoc Warnings [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:60: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:61: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:63: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:69: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:87: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:88: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:90: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:95: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:96: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Block.java:67: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Configuration.java:275: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DAO.java:57: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DataType.java:346: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:66: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:67: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag > ```
1.0
Fix Javadoc warnings - The build currently complains about the following Javadoc warnings: ``` [WARNING] Javadoc Warnings [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:60: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:61: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:63: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:69: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:87: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:88: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:90: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:95: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\BatchBindStep.java:96: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Block.java:67: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\CommonTableExpression.java:51: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Configuration.java:275: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Converter.java:197: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:42: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:54: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\ConverterProvider.java:55: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DAO.java:57: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DataType.java:346: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:66: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\Declaration.java:67: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList1.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList10.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList11.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList12.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList13.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList14.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList15.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList16.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList17.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList18.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList19.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList2.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList20.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList21.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList22.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList3.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList4.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList5.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList6.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList7.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList8.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DerivedColumnList9.java:68: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:907: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:934: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:964: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1032: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1064: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1099: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1172: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1202: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1235: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1301: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1328: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1358: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1426: warning - invalid usage of tag > [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag < [WARNING] C:\Users\lukas\workspace-jooq-pro\jOOQ\jOOQ\src\main\java\org\jooq\DSLContext.java:1460: warning - invalid usage of tag > ```
non_process
fix javadoc warnings the build currently complains about the following javadoc warnings javadoc warnings c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq batchbindstep java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq block java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq commontableexpression java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq commontableexpression java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq configuration java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converter java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converter java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq converterprovider java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dao java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq datatype java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq declaration java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq declaration java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq derivedcolumnlist java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq derivedcolumnlist java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag c users lukas workspace jooq pro jooq jooq src main java org jooq dslcontext java warning invalid usage of tag
0
9,109
12,192,306,098
IssuesEvent
2020-04-29 12:44:07
naoki-shigehisa/paper
https://api.github.com/repos/naoki-shigehisa/paper
opened
Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems
2019 Gaussian Process recommendation
## 0. 論文 タイトル:[Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems](https://arxiv.org/abs/1912.09593) 著者: ![スクリーンショット 2020-04-29 21 43 43](https://user-images.githubusercontent.com/43877096/80597234-84f07280-8a62-11ea-9b54-f36beb06da0f.png) arXiv投稿日:2019/12/19 学会/ジャーナル: ## 1. どんなもの? ## 2. 先行研究と比べてどこがすごい? ## 3. 技術や手法のキモはどこ? ## 4. どうやって有効だと検証した? ## 5. 議論はある? ## 6. 次に読むべき論文は?
1.0
Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems - ## 0. 論文 タイトル:[Gaussian Process Latent Variable Model Factorization for Context-aware Recommender Systems](https://arxiv.org/abs/1912.09593) 著者: ![スクリーンショット 2020-04-29 21 43 43](https://user-images.githubusercontent.com/43877096/80597234-84f07280-8a62-11ea-9b54-f36beb06da0f.png) arXiv投稿日:2019/12/19 学会/ジャーナル: ## 1. どんなもの? ## 2. 先行研究と比べてどこがすごい? ## 3. 技術や手法のキモはどこ? ## 4. どうやって有効だと検証した? ## 5. 議論はある? ## 6. 次に読むべき論文は?
process
gaussian process latent variable model factorization for context aware recommender systems 論文 タイトル: 著者: arxiv投稿日: 学会 ジャーナル: どんなもの? 先行研究と比べてどこがすごい? 技術や手法のキモはどこ? どうやって有効だと検証した? 議論はある? 次に読むべき論文は?
1
20,243
26,861,732,740
IssuesEvent
2023-02-03 19:03:22
python/cpython
https://api.github.com/repos/python/cpython
reopened
multiprocessing's default posix start method of `'fork'` is broken: change to `'spawn'`
type-feature 3.12 expert-multiprocessing
BPO | [40379](https://bugs.python.org/issue40379) --- | :--- Nosy | @pitrou, @mgorny, @Julian, @wimglenn, @applio, @itamarst <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-04-24.18:22:23.389> labels = ['3.8', 'type-bug', '3.7', '3.9'] title = "multiprocessing's default start method of fork()-without-exec() is broken" updated_at = <Date 2022-02-11.16:13:53.872> user = 'https://bugs.python.org/itamarst' ``` bugs.python.org fields: ```python activity = <Date 2022-02-11.16:13:53.872> actor = 'mgorny' assignee = 'none' closed = False closed_date = None closer = None components = [] creation = <Date 2020-04-24.18:22:23.389> creator = 'itamarst' dependencies = [] files = [] hgrepos = [] issue_num = 40379 keywords = [] message_count = 11.0 messages = ['367210', '367211', '368173', '380478', '392358', '392501', '392503', '392506', '392507', '392508', '413081'] nosy_count = 8.0 nosy_names = ['pitrou', 'mgorny', 'Julian', 'wim.glenn', 'itamarst', 'davin', 'itamarst2', 'aduncan'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue40379' versions = ['Python 3.5', 'Python 3.6', 'Python 3.7', 'Python 3.8', 'Python 3.9'] ``` </p></details> <!-- gh-linked-prs --> ### Linked PRs * gh-100618 * gh-101530 <!-- /gh-linked-prs -->
1.0
multiprocessing's default posix start method of `'fork'` is broken: change to `'spawn'` - BPO | [40379](https://bugs.python.org/issue40379) --- | :--- Nosy | @pitrou, @mgorny, @Julian, @wimglenn, @applio, @itamarst <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-04-24.18:22:23.389> labels = ['3.8', 'type-bug', '3.7', '3.9'] title = "multiprocessing's default start method of fork()-without-exec() is broken" updated_at = <Date 2022-02-11.16:13:53.872> user = 'https://bugs.python.org/itamarst' ``` bugs.python.org fields: ```python activity = <Date 2022-02-11.16:13:53.872> actor = 'mgorny' assignee = 'none' closed = False closed_date = None closer = None components = [] creation = <Date 2020-04-24.18:22:23.389> creator = 'itamarst' dependencies = [] files = [] hgrepos = [] issue_num = 40379 keywords = [] message_count = 11.0 messages = ['367210', '367211', '368173', '380478', '392358', '392501', '392503', '392506', '392507', '392508', '413081'] nosy_count = 8.0 nosy_names = ['pitrou', 'mgorny', 'Julian', 'wim.glenn', 'itamarst', 'davin', 'itamarst2', 'aduncan'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue40379' versions = ['Python 3.5', 'Python 3.6', 'Python 3.7', 'Python 3.8', 'Python 3.9'] ``` </p></details> <!-- gh-linked-prs --> ### Linked PRs * gh-100618 * gh-101530 <!-- /gh-linked-prs -->
process
multiprocessing s default posix start method of fork is broken change to spawn bpo nosy pitrou mgorny julian wimglenn applio itamarst note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title multiprocessing s default start method of fork without exec is broken updated at user bugs python org fields python activity actor mgorny assignee none closed false closed date none closer none components creation creator itamarst dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type behavior url versions linked prs gh gh
1
497,319
14,367,968,311
IssuesEvent
2020-12-01 07:38:16
teamforus/forus
https://api.github.com/repos/teamforus/forus
closed
CR: Webshop - remove "Hoe het werkt" below header description
Priority: Must have
@maxvisser commented on [Tue Sep 15 2020](https://github.com/teamforus/general/issues/445) Learn more about change requests here: https://bit.ly/39CWeEE ### Requested by: - Geertruidenberg ### Change description I don't think such a button really fits there if there is a menu at the top of the page. The button 'how it works' and 'explanation' in the menu link to the same page. I would delete "Hoe werkt het". ![image](https://user-images.githubusercontent.com/10818702/98953111-8d8b3080-24fc-11eb-9129-52ca7cc1af6b.png)
1.0
CR: Webshop - remove "Hoe het werkt" below header description - @maxvisser commented on [Tue Sep 15 2020](https://github.com/teamforus/general/issues/445) Learn more about change requests here: https://bit.ly/39CWeEE ### Requested by: - Geertruidenberg ### Change description I don't think such a button really fits there if there is a menu at the top of the page. The button 'how it works' and 'explanation' in the menu link to the same page. I would delete "Hoe werkt het". ![image](https://user-images.githubusercontent.com/10818702/98953111-8d8b3080-24fc-11eb-9129-52ca7cc1af6b.png)
non_process
cr webshop remove hoe het werkt below header description maxvisser commented on learn more about change requests here requested by geertruidenberg change description i don t think such a button really fits there if there is a menu at the top of the page the button how it works and explanation in the menu link to the same page i would delete hoe werkt het
0
22,568
31,790,020,810
IssuesEvent
2023-09-13 02:00:10
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Wed, 13 Sep 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### SoccerNet 2023 Challenges Results - **Authors:** Anthony Cioppa, Silvio Giancola, Vladimir Somers, Floriane Magera, Xin Zhou, Hassan Mkhallati, Adrien Deliège, Jan Held, Carlos Hinojosa, Amir M. Mansourian, Pierre Miralles, Olivier Barnich, Christophe De Vleeschouwer, Alexandre Alahi, Bernard Ghanem, Marc Van Droogenbroeck, Abdullah Kamal, Adrien Maglo, Albert Clapés, Amr Abdelaziz, Artur Xarles, Astrid Orcesi, Atom Scott, Bin Liu, Byoungkwon Lim, Chen Chen, Fabian Deuser, Feng Yan, Fufu Yu, Gal Shitrit, Guanshuo Wang, Gyusik Choi, Hankyul Kim, Hao Guo, Hasby Fahrudin, Hidenari Koguchi, Håkan Ardö, Ibrahim Salah, Ido Yerushalmy, Iftikar Muhammad, Ikuma Uchida, Ishay Be'ery, Jaonary Rabarisoa, Jeongae Lee, Jiajun Fu, Jianqin Yin, Jinghang Xu, Jongho Nang, Julien Denize, Junjie Li, Junpei Zhang, Juntae Kim, Kamil Synowiec, et al. (49 additional authors not shown) - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2309.06006 - **Pdf link:** https://arxiv.org/pdf/2309.06006 - **Abstract** The SoccerNet 2023 challenges were the third annual video understanding challenges organized by the SoccerNet team. For this third edition, the challenges were composed of seven vision-based tasks split into three main themes. The first theme, broadcast video understanding, is composed of three high-level tasks related to describing events occurring in the video broadcasts: (1) action spotting, focusing on retrieving all timestamps related to global actions in soccer, (2) ball action spotting, focusing on retrieving all timestamps related to the soccer ball change of state, and (3) dense video captioning, focusing on describing the broadcast with natural language and anchored timestamps. The second theme, field understanding, relates to the single task of (4) camera calibration, focusing on retrieving the intrinsic and extrinsic camera parameters from images. The third and last theme, player understanding, is composed of three low-level tasks related to extracting information about the players: (5) re-identification, focusing on retrieving the same players across multiple views, (6) multiple object tracking, focusing on tracking players and the ball through unedited video streams, and (7) jersey number recognition, focusing on recognizing the jersey number of players from tracklets. Compared to the previous editions of the SoccerNet challenges, tasks (2-3-7) are novel, including new annotations and data, task (4) was enhanced with more data and annotations, and task (6) now focuses on end-to-end approaches. More information on the tasks, challenges, and leaderboards are available on https://www.soccer-net.org. Baselines and development kits can be found on https://github.com/SoccerNet. ### JOADAA: joint online action detection and action anticipation - **Authors:** Mohammed Guermal, Francois Bremond, Rui Dai, Abid Ali - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2309.06130 - **Pdf link:** https://arxiv.org/pdf/2309.06130 - **Abstract** Action anticipation involves forecasting future actions by connecting past events to future ones. However, this reasoning ignores the real-life hierarchy of events which is considered to be composed of three main parts: past, present, and future. We argue that considering these three main parts and their dependencies could improve performance. On the other hand, online action detection is the task of predicting actions in a streaming manner. In this case, one has access only to the past and present information. Therefore, in online action detection (OAD) the existing approaches miss semantics or future information which limits their performance. To sum up, for both of these tasks, the complete set of knowledge (past-present-future) is missing, which makes it challenging to infer action dependencies, therefore having low performances. To address this limitation, we propose to fuse both tasks into a single uniform architecture. By combining action anticipation and online action detection, our approach can cover the missing dependencies of future information in online action detection. This method referred to as JOADAA, presents a uniform model that jointly performs action anticipation and online action detection. We validate our proposed model on three challenging datasets: THUMOS'14, which is a sparsely annotated dataset with one action per time step, CHARADES, and Multi-THUMOS, two densely annotated datasets with more complex scenarios. JOADAA achieves SOTA results on these benchmarks for both tasks. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Rice Plant Disease Detection and Diagnosis using Deep Convolutional Neural Networks and Multispectral Imaging - **Authors:** Yara Ali Alnaggar, Ahmad Sebaq, Karim Amer, ElSayed Naeem, Mohamed Elhelw - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2309.05818 - **Pdf link:** https://arxiv.org/pdf/2309.05818 - **Abstract** Rice is considered a strategic crop in Egypt as it is regularly consumed in the Egyptian people's diet. Even though Egypt is the highest rice producer in Africa with a share of 6 million tons per year, it still imports rice to satisfy its local needs due to production loss, especially due to rice disease. Rice blast disease is responsible for 30% loss in rice production worldwide. Therefore, it is crucial to target limiting yield damage by detecting rice crops diseases in its early stages. This paper introduces a public multispectral and RGB images dataset and a deep learning pipeline for rice plant disease detection using multi-modal data. The collected multispectral images consist of Red, Green and Near-Infrared channels and we show that using multispectral along with RGB channels as input archives a higher F1 accuracy compared to using RGB input only. ### Combining deep learning and street view imagery to map smallholder crop types - **Authors:** Jordi Laguarta, Thomas Friedel, Sherrie Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2309.05930 - **Pdf link:** https://arxiv.org/pdf/2309.05930 - **Abstract** Accurate crop type maps are an essential source of information for monitoring yield progress at scale, projecting global crop production, and planning effective policies. To date, however, crop type maps remain challenging to create in low and middle-income countries due to a lack of ground truth labels for training machine learning models. Field surveys are the gold standard in terms of accuracy but require an often-prohibitively large amount of time, money, and statistical capacity. In recent years, street-level imagery, such as Google Street View, KartaView, and Mapillary, has become available around the world. Such imagery contains rich information about crop types grown at particular locations and times. In this work, we develop an automated system to generate crop type ground references using deep learning and Google Street View imagery. The method efficiently curates a set of street view images containing crop fields, trains a model to predict crop type by utilizing weakly-labelled images from disparate out-of-domain sources, and combines predicted labels with remote sensing time series to create a wall-to-wall crop type map. We show that, in Thailand, the resulting country-wide map of rice, cassava, maize, and sugarcane achieves an accuracy of 93%. As the availability of roadside imagery expands, our pipeline provides a way to map crop types at scale around the globe, especially in underserved smallholder regions. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Real-Time Semantic Segmentation: A Brief Survey & Comparative Study in Remote Sensing - **Authors:** Clifford Broni-Bediako, Junshi Xia, Naoto Yokoya - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.06047 - **Pdf link:** https://arxiv.org/pdf/2309.06047 - **Abstract** Real-time semantic segmentation of remote sensing imagery is a challenging task that requires a trade-off between effectiveness and efficiency. It has many applications including tracking forest fires, detecting changes in land use and land cover, crop health monitoring, and so on. With the success of efficient deep learning methods (i.e., efficient deep neural networks) for real-time semantic segmentation in computer vision, researchers have adopted these efficient deep neural networks in remote sensing image analysis. This paper begins with a summary of the fundamental compression methods for designing efficient deep neural networks and provides a brief but comprehensive survey, outlining the recent developments in real-time semantic segmentation of remote sensing imagery. We examine several seminal efficient deep learning methods, placing them in a taxonomy based on the network architecture design approach. Furthermore, we evaluate the quality and efficiency of some existing efficient deep neural networks on a publicly available remote sensing semantic segmentation benchmark dataset, the OpenEarthMap. The experimental results of an extensive comparative study demonstrate that most of the existing efficient deep neural networks have good segmentation quality, but they suffer low inference speed (i.e., high latency rate), which may limit their capability of deployment in real-time applications of remote sensing image segmentation. We provide some insights into the current trend and future research directions for real-time semantic segmentation of remote sensing imagery. ## Keyword: RAW ### Adversarial Attacks Assessment of Salient Object Detection via Symbolic Learning - **Authors:** Gustavo Olague, Roberto Pineda, Gerardo Ibarra-Vazquez, Matthieu Olague, Axel Martinez, Sambit Bakshi, Jonathan Vargas, Isnardo Reducindo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2309.05900 - **Pdf link:** https://arxiv.org/pdf/2309.05900 - **Abstract** Machine learning is at the center of mainstream technology and outperforms classical approaches to handcrafted feature design. Aside from its learning process for artificial feature extraction, it has an end-to-end paradigm from input to output, reaching outstandingly accurate results. However, security concerns about its robustness to malicious and imperceptible perturbations have drawn attention since its prediction can be changed entirely. Salient object detection is a research area where deep convolutional neural networks have proven effective but whose trustworthiness represents a significant issue requiring analysis and solutions to hackers' attacks. Brain programming is a kind of symbolic learning in the vein of good old-fashioned artificial intelligence. This work provides evidence that symbolic learning robustness is crucial in designing reliable visual attention systems since it can withstand even the most intense perturbations. We test this evolutionary computation methodology against several adversarial attacks and noise perturbations using standard databases and a real-world problem of a shorebird called the Snowy Plover portraying a visual attention task. We compare our methodology with five different deep learning approaches, proving that they do not match the symbolic paradigm regarding robustness. All neural networks suffer significant performance losses, while brain programming stands its ground and remains unaffected. Also, by studying the Snowy Plover, we remark on the importance of security in surveillance activities regarding wildlife protection and conservation. ### TSSAT: Two-Stage Statistics-Aware Transformation for Artistic Style Transfer - **Authors:** Haibo Chen, Lei Zhao, Jun Li, Jian Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.06004 - **Pdf link:** https://arxiv.org/pdf/2309.06004 - **Abstract** Artistic style transfer aims to create new artistic images by rendering a given photograph with the target artistic style. Existing methods learn styles simply based on global statistics or local patches, lacking careful consideration of the drawing process in practice. Consequently, the stylization results either fail to capture abundant and diversified local style patterns, or contain undesired semantic information of the style image and deviate from the global style distribution. To address this issue, we imitate the drawing process of humans and propose a Two-Stage Statistics-Aware Transformation (TSSAT) module, which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics (instead of local features) in a patch-wise manner, significantly improving the stylization effects. Moreover, to further enhance both content and style representations, we introduce two novel losses: an attention-based content loss and a patch-based style loss, where the former enables better content preservation by enforcing the semantic relation in the content image to be retained during stylization, and the latter focuses on increasing the local style similarity between the style and stylized images. Extensive qualitative and quantitative experiments verify the effectiveness of our method. ## Keyword: raw image There is no result
2.0
New submissions for Wed, 13 Sep 23 - ## Keyword: events ### SoccerNet 2023 Challenges Results - **Authors:** Anthony Cioppa, Silvio Giancola, Vladimir Somers, Floriane Magera, Xin Zhou, Hassan Mkhallati, Adrien Deliège, Jan Held, Carlos Hinojosa, Amir M. Mansourian, Pierre Miralles, Olivier Barnich, Christophe De Vleeschouwer, Alexandre Alahi, Bernard Ghanem, Marc Van Droogenbroeck, Abdullah Kamal, Adrien Maglo, Albert Clapés, Amr Abdelaziz, Artur Xarles, Astrid Orcesi, Atom Scott, Bin Liu, Byoungkwon Lim, Chen Chen, Fabian Deuser, Feng Yan, Fufu Yu, Gal Shitrit, Guanshuo Wang, Gyusik Choi, Hankyul Kim, Hao Guo, Hasby Fahrudin, Hidenari Koguchi, Håkan Ardö, Ibrahim Salah, Ido Yerushalmy, Iftikar Muhammad, Ikuma Uchida, Ishay Be'ery, Jaonary Rabarisoa, Jeongae Lee, Jiajun Fu, Jianqin Yin, Jinghang Xu, Jongho Nang, Julien Denize, Junjie Li, Junpei Zhang, Juntae Kim, Kamil Synowiec, et al. (49 additional authors not shown) - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2309.06006 - **Pdf link:** https://arxiv.org/pdf/2309.06006 - **Abstract** The SoccerNet 2023 challenges were the third annual video understanding challenges organized by the SoccerNet team. For this third edition, the challenges were composed of seven vision-based tasks split into three main themes. The first theme, broadcast video understanding, is composed of three high-level tasks related to describing events occurring in the video broadcasts: (1) action spotting, focusing on retrieving all timestamps related to global actions in soccer, (2) ball action spotting, focusing on retrieving all timestamps related to the soccer ball change of state, and (3) dense video captioning, focusing on describing the broadcast with natural language and anchored timestamps. The second theme, field understanding, relates to the single task of (4) camera calibration, focusing on retrieving the intrinsic and extrinsic camera parameters from images. The third and last theme, player understanding, is composed of three low-level tasks related to extracting information about the players: (5) re-identification, focusing on retrieving the same players across multiple views, (6) multiple object tracking, focusing on tracking players and the ball through unedited video streams, and (7) jersey number recognition, focusing on recognizing the jersey number of players from tracklets. Compared to the previous editions of the SoccerNet challenges, tasks (2-3-7) are novel, including new annotations and data, task (4) was enhanced with more data and annotations, and task (6) now focuses on end-to-end approaches. More information on the tasks, challenges, and leaderboards are available on https://www.soccer-net.org. Baselines and development kits can be found on https://github.com/SoccerNet. ### JOADAA: joint online action detection and action anticipation - **Authors:** Mohammed Guermal, Francois Bremond, Rui Dai, Abid Ali - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2309.06130 - **Pdf link:** https://arxiv.org/pdf/2309.06130 - **Abstract** Action anticipation involves forecasting future actions by connecting past events to future ones. However, this reasoning ignores the real-life hierarchy of events which is considered to be composed of three main parts: past, present, and future. We argue that considering these three main parts and their dependencies could improve performance. On the other hand, online action detection is the task of predicting actions in a streaming manner. In this case, one has access only to the past and present information. Therefore, in online action detection (OAD) the existing approaches miss semantics or future information which limits their performance. To sum up, for both of these tasks, the complete set of knowledge (past-present-future) is missing, which makes it challenging to infer action dependencies, therefore having low performances. To address this limitation, we propose to fuse both tasks into a single uniform architecture. By combining action anticipation and online action detection, our approach can cover the missing dependencies of future information in online action detection. This method referred to as JOADAA, presents a uniform model that jointly performs action anticipation and online action detection. We validate our proposed model on three challenging datasets: THUMOS'14, which is a sparsely annotated dataset with one action per time step, CHARADES, and Multi-THUMOS, two densely annotated datasets with more complex scenarios. JOADAA achieves SOTA results on these benchmarks for both tasks. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Rice Plant Disease Detection and Diagnosis using Deep Convolutional Neural Networks and Multispectral Imaging - **Authors:** Yara Ali Alnaggar, Ahmad Sebaq, Karim Amer, ElSayed Naeem, Mohamed Elhelw - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2309.05818 - **Pdf link:** https://arxiv.org/pdf/2309.05818 - **Abstract** Rice is considered a strategic crop in Egypt as it is regularly consumed in the Egyptian people's diet. Even though Egypt is the highest rice producer in Africa with a share of 6 million tons per year, it still imports rice to satisfy its local needs due to production loss, especially due to rice disease. Rice blast disease is responsible for 30% loss in rice production worldwide. Therefore, it is crucial to target limiting yield damage by detecting rice crops diseases in its early stages. This paper introduces a public multispectral and RGB images dataset and a deep learning pipeline for rice plant disease detection using multi-modal data. The collected multispectral images consist of Red, Green and Near-Infrared channels and we show that using multispectral along with RGB channels as input archives a higher F1 accuracy compared to using RGB input only. ### Combining deep learning and street view imagery to map smallholder crop types - **Authors:** Jordi Laguarta, Thomas Friedel, Sherrie Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2309.05930 - **Pdf link:** https://arxiv.org/pdf/2309.05930 - **Abstract** Accurate crop type maps are an essential source of information for monitoring yield progress at scale, projecting global crop production, and planning effective policies. To date, however, crop type maps remain challenging to create in low and middle-income countries due to a lack of ground truth labels for training machine learning models. Field surveys are the gold standard in terms of accuracy but require an often-prohibitively large amount of time, money, and statistical capacity. In recent years, street-level imagery, such as Google Street View, KartaView, and Mapillary, has become available around the world. Such imagery contains rich information about crop types grown at particular locations and times. In this work, we develop an automated system to generate crop type ground references using deep learning and Google Street View imagery. The method efficiently curates a set of street view images containing crop fields, trains a model to predict crop type by utilizing weakly-labelled images from disparate out-of-domain sources, and combines predicted labels with remote sensing time series to create a wall-to-wall crop type map. We show that, in Thailand, the resulting country-wide map of rice, cassava, maize, and sugarcane achieves an accuracy of 93%. As the availability of roadside imagery expands, our pipeline provides a way to map crop types at scale around the globe, especially in underserved smallholder regions. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Real-Time Semantic Segmentation: A Brief Survey & Comparative Study in Remote Sensing - **Authors:** Clifford Broni-Bediako, Junshi Xia, Naoto Yokoya - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.06047 - **Pdf link:** https://arxiv.org/pdf/2309.06047 - **Abstract** Real-time semantic segmentation of remote sensing imagery is a challenging task that requires a trade-off between effectiveness and efficiency. It has many applications including tracking forest fires, detecting changes in land use and land cover, crop health monitoring, and so on. With the success of efficient deep learning methods (i.e., efficient deep neural networks) for real-time semantic segmentation in computer vision, researchers have adopted these efficient deep neural networks in remote sensing image analysis. This paper begins with a summary of the fundamental compression methods for designing efficient deep neural networks and provides a brief but comprehensive survey, outlining the recent developments in real-time semantic segmentation of remote sensing imagery. We examine several seminal efficient deep learning methods, placing them in a taxonomy based on the network architecture design approach. Furthermore, we evaluate the quality and efficiency of some existing efficient deep neural networks on a publicly available remote sensing semantic segmentation benchmark dataset, the OpenEarthMap. The experimental results of an extensive comparative study demonstrate that most of the existing efficient deep neural networks have good segmentation quality, but they suffer low inference speed (i.e., high latency rate), which may limit their capability of deployment in real-time applications of remote sensing image segmentation. We provide some insights into the current trend and future research directions for real-time semantic segmentation of remote sensing imagery. ## Keyword: RAW ### Adversarial Attacks Assessment of Salient Object Detection via Symbolic Learning - **Authors:** Gustavo Olague, Roberto Pineda, Gerardo Ibarra-Vazquez, Matthieu Olague, Axel Martinez, Sambit Bakshi, Jonathan Vargas, Isnardo Reducindo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2309.05900 - **Pdf link:** https://arxiv.org/pdf/2309.05900 - **Abstract** Machine learning is at the center of mainstream technology and outperforms classical approaches to handcrafted feature design. Aside from its learning process for artificial feature extraction, it has an end-to-end paradigm from input to output, reaching outstandingly accurate results. However, security concerns about its robustness to malicious and imperceptible perturbations have drawn attention since its prediction can be changed entirely. Salient object detection is a research area where deep convolutional neural networks have proven effective but whose trustworthiness represents a significant issue requiring analysis and solutions to hackers' attacks. Brain programming is a kind of symbolic learning in the vein of good old-fashioned artificial intelligence. This work provides evidence that symbolic learning robustness is crucial in designing reliable visual attention systems since it can withstand even the most intense perturbations. We test this evolutionary computation methodology against several adversarial attacks and noise perturbations using standard databases and a real-world problem of a shorebird called the Snowy Plover portraying a visual attention task. We compare our methodology with five different deep learning approaches, proving that they do not match the symbolic paradigm regarding robustness. All neural networks suffer significant performance losses, while brain programming stands its ground and remains unaffected. Also, by studying the Snowy Plover, we remark on the importance of security in surveillance activities regarding wildlife protection and conservation. ### TSSAT: Two-Stage Statistics-Aware Transformation for Artistic Style Transfer - **Authors:** Haibo Chen, Lei Zhao, Jun Li, Jian Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.06004 - **Pdf link:** https://arxiv.org/pdf/2309.06004 - **Abstract** Artistic style transfer aims to create new artistic images by rendering a given photograph with the target artistic style. Existing methods learn styles simply based on global statistics or local patches, lacking careful consideration of the drawing process in practice. Consequently, the stylization results either fail to capture abundant and diversified local style patterns, or contain undesired semantic information of the style image and deviate from the global style distribution. To address this issue, we imitate the drawing process of humans and propose a Two-Stage Statistics-Aware Transformation (TSSAT) module, which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics (instead of local features) in a patch-wise manner, significantly improving the stylization effects. Moreover, to further enhance both content and style representations, we introduce two novel losses: an attention-based content loss and a patch-based style loss, where the former enables better content preservation by enforcing the semantic relation in the content image to be retained during stylization, and the latter focuses on increasing the local style similarity between the style and stylized images. Extensive qualitative and quantitative experiments verify the effectiveness of our method. ## Keyword: raw image There is no result
process
new submissions for wed sep keyword events soccernet challenges results authors anthony cioppa silvio giancola vladimir somers floriane magera xin zhou hassan mkhallati adrien deliège jan held carlos hinojosa amir m mansourian pierre miralles olivier barnich christophe de vleeschouwer alexandre alahi bernard ghanem marc van droogenbroeck abdullah kamal adrien maglo albert clapés amr abdelaziz artur xarles astrid orcesi atom scott bin liu byoungkwon lim chen chen fabian deuser feng yan fufu yu gal shitrit guanshuo wang gyusik choi hankyul kim hao guo hasby fahrudin hidenari koguchi håkan ardö ibrahim salah ido yerushalmy iftikar muhammad ikuma uchida ishay be ery jaonary rabarisoa jeongae lee jiajun fu jianqin yin jinghang xu jongho nang julien denize junjie li junpei zhang juntae kim kamil synowiec et al additional authors not shown subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract the soccernet challenges were the third annual video understanding challenges organized by the soccernet team for this third edition the challenges were composed of seven vision based tasks split into three main themes the first theme broadcast video understanding is composed of three high level tasks related to describing events occurring in the video broadcasts action spotting focusing on retrieving all timestamps related to global actions in soccer ball action spotting focusing on retrieving all timestamps related to the soccer ball change of state and dense video captioning focusing on describing the broadcast with natural language and anchored timestamps the second theme field understanding relates to the single task of camera calibration focusing on retrieving the intrinsic and extrinsic camera parameters from images the third and last theme player understanding is composed of three low level tasks related to extracting information about the players re identification focusing on retrieving the same players across multiple views multiple object tracking focusing on tracking players and the ball through unedited video streams and jersey number recognition focusing on recognizing the jersey number of players from tracklets compared to the previous editions of the soccernet challenges tasks are novel including new annotations and data task was enhanced with more data and annotations and task now focuses on end to end approaches more information on the tasks challenges and leaderboards are available on baselines and development kits can be found on joadaa joint online action detection and action anticipation authors mohammed guermal francois bremond rui dai abid ali subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract action anticipation involves forecasting future actions by connecting past events to future ones however this reasoning ignores the real life hierarchy of events which is considered to be composed of three main parts past present and future we argue that considering these three main parts and their dependencies could improve performance on the other hand online action detection is the task of predicting actions in a streaming manner in this case one has access only to the past and present information therefore in online action detection oad the existing approaches miss semantics or future information which limits their performance to sum up for both of these tasks the complete set of knowledge past present future is missing which makes it challenging to infer action dependencies therefore having low performances to address this limitation we propose to fuse both tasks into a single uniform architecture by combining action anticipation and online action detection our approach can cover the missing dependencies of future information in online action detection this method referred to as joadaa presents a uniform model that jointly performs action anticipation and online action detection we validate our proposed model on three challenging datasets thumos which is a sparsely annotated dataset with one action per time step charades and multi thumos two densely annotated datasets with more complex scenarios joadaa achieves sota results on these benchmarks for both tasks keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp rice plant disease detection and diagnosis using deep convolutional neural networks and multispectral imaging authors yara ali alnaggar ahmad sebaq karim amer elsayed naeem mohamed elhelw subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract rice is considered a strategic crop in egypt as it is regularly consumed in the egyptian people s diet even though egypt is the highest rice producer in africa with a share of million tons per year it still imports rice to satisfy its local needs due to production loss especially due to rice disease rice blast disease is responsible for loss in rice production worldwide therefore it is crucial to target limiting yield damage by detecting rice crops diseases in its early stages this paper introduces a public multispectral and rgb images dataset and a deep learning pipeline for rice plant disease detection using multi modal data the collected multispectral images consist of red green and near infrared channels and we show that using multispectral along with rgb channels as input archives a higher accuracy compared to using rgb input only combining deep learning and street view imagery to map smallholder crop types authors jordi laguarta thomas friedel sherrie wang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract accurate crop type maps are an essential source of information for monitoring yield progress at scale projecting global crop production and planning effective policies to date however crop type maps remain challenging to create in low and middle income countries due to a lack of ground truth labels for training machine learning models field surveys are the gold standard in terms of accuracy but require an often prohibitively large amount of time money and statistical capacity in recent years street level imagery such as google street view kartaview and mapillary has become available around the world such imagery contains rich information about crop types grown at particular locations and times in this work we develop an automated system to generate crop type ground references using deep learning and google street view imagery the method efficiently curates a set of street view images containing crop fields trains a model to predict crop type by utilizing weakly labelled images from disparate out of domain sources and combines predicted labels with remote sensing time series to create a wall to wall crop type map we show that in thailand the resulting country wide map of rice cassava maize and sugarcane achieves an accuracy of as the availability of roadside imagery expands our pipeline provides a way to map crop types at scale around the globe especially in underserved smallholder regions keyword image signal processing there is no result keyword image signal process there is no result keyword compression real time semantic segmentation a brief survey comparative study in remote sensing authors clifford broni bediako junshi xia naoto yokoya subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract real time semantic segmentation of remote sensing imagery is a challenging task that requires a trade off between effectiveness and efficiency it has many applications including tracking forest fires detecting changes in land use and land cover crop health monitoring and so on with the success of efficient deep learning methods i e efficient deep neural networks for real time semantic segmentation in computer vision researchers have adopted these efficient deep neural networks in remote sensing image analysis this paper begins with a summary of the fundamental compression methods for designing efficient deep neural networks and provides a brief but comprehensive survey outlining the recent developments in real time semantic segmentation of remote sensing imagery we examine several seminal efficient deep learning methods placing them in a taxonomy based on the network architecture design approach furthermore we evaluate the quality and efficiency of some existing efficient deep neural networks on a publicly available remote sensing semantic segmentation benchmark dataset the openearthmap the experimental results of an extensive comparative study demonstrate that most of the existing efficient deep neural networks have good segmentation quality but they suffer low inference speed i e high latency rate which may limit their capability of deployment in real time applications of remote sensing image segmentation we provide some insights into the current trend and future research directions for real time semantic segmentation of remote sensing imagery keyword raw adversarial attacks assessment of salient object detection via symbolic learning authors gustavo olague roberto pineda gerardo ibarra vazquez matthieu olague axel martinez sambit bakshi jonathan vargas isnardo reducindo subjects computer vision and pattern recognition cs cv cryptography and security cs cr machine learning cs lg neural and evolutionary computing cs ne arxiv link pdf link abstract machine learning is at the center of mainstream technology and outperforms classical approaches to handcrafted feature design aside from its learning process for artificial feature extraction it has an end to end paradigm from input to output reaching outstandingly accurate results however security concerns about its robustness to malicious and imperceptible perturbations have drawn attention since its prediction can be changed entirely salient object detection is a research area where deep convolutional neural networks have proven effective but whose trustworthiness represents a significant issue requiring analysis and solutions to hackers attacks brain programming is a kind of symbolic learning in the vein of good old fashioned artificial intelligence this work provides evidence that symbolic learning robustness is crucial in designing reliable visual attention systems since it can withstand even the most intense perturbations we test this evolutionary computation methodology against several adversarial attacks and noise perturbations using standard databases and a real world problem of a shorebird called the snowy plover portraying a visual attention task we compare our methodology with five different deep learning approaches proving that they do not match the symbolic paradigm regarding robustness all neural networks suffer significant performance losses while brain programming stands its ground and remains unaffected also by studying the snowy plover we remark on the importance of security in surveillance activities regarding wildlife protection and conservation tssat two stage statistics aware transformation for artistic style transfer authors haibo chen lei zhao jun li jian yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract artistic style transfer aims to create new artistic images by rendering a given photograph with the target artistic style existing methods learn styles simply based on global statistics or local patches lacking careful consideration of the drawing process in practice consequently the stylization results either fail to capture abundant and diversified local style patterns or contain undesired semantic information of the style image and deviate from the global style distribution to address this issue we imitate the drawing process of humans and propose a two stage statistics aware transformation tssat module which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics instead of local features in a patch wise manner significantly improving the stylization effects moreover to further enhance both content and style representations we introduce two novel losses an attention based content loss and a patch based style loss where the former enables better content preservation by enforcing the semantic relation in the content image to be retained during stylization and the latter focuses on increasing the local style similarity between the style and stylized images extensive qualitative and quantitative experiments verify the effectiveness of our method keyword raw image there is no result
1
14,645
17,773,567,497
IssuesEvent
2021-08-30 16:16:47
googleapis/python-storage
https://api.github.com/repos/googleapis/python-storage
closed
Dependency Dashboard
api: storage type: process
This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/googleapis-common-protos-1.x -->[chore(deps): update dependency googleapis-common-protos to <1.53.1](../pull/432) - [ ] <!-- recreate-branch=renovate/google-api-core-2.x -->[chore(deps): update dependency google-api-core to v2](../pull/558) - [ ] <!-- recreate-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2](../pull/542) - [ ] <!-- recreate-branch=renovate/google-resumable-media-2.x -->[chore(deps): update dependency google-resumable-media to v2](../pull/547) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/googleapis-common-protos-1.x -->[chore(deps): update dependency googleapis-common-protos to <1.53.1](../pull/432) - [ ] <!-- recreate-branch=renovate/google-api-core-2.x -->[chore(deps): update dependency google-api-core to v2](../pull/558) - [ ] <!-- recreate-branch=renovate/google-auth-2.x -->[chore(deps): update dependency google-auth to v2](../pull/542) - [ ] <!-- recreate-branch=renovate/google-resumable-media-2.x -->[chore(deps): update dependency google-resumable-media to v2](../pull/547) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue provides visibility into renovate updates and their statuses ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull pull pull check this box to trigger a request for renovate to run again on this repository
1
119,972
17,644,003,757
IssuesEvent
2021-08-20 01:26:19
AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
https://api.github.com/repos/AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
opened
CVE-2021-29529 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2021-29529 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /FinalProject/requirements.txt</p> <p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. An attacker can trigger a heap buffer overflow in `tf.raw_ops.QuantizedResizeBilinear` by manipulating input values so that float rounding results in off-by-one error in accessing image elements. This is because the implementation(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L62-L66) computes two integers (representing the upper and lower bounds for interpolation) by ceiling and flooring a floating point value. For some values of `in`, `interpolation->upper[i]` might be smaller than `interpolation->lower[i]`. This is an issue if `interpolation->upper[i]` is capped at `in_size-1` as it means that `interpolation->lower[i]` points outside of the image. Then, in the interpolation code(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L245-L264), this would result in heap buffer overflow. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29529>CVE-2021-29529</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29529 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-29529 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /FinalProject/requirements.txt</p> <p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. An attacker can trigger a heap buffer overflow in `tf.raw_ops.QuantizedResizeBilinear` by manipulating input values so that float rounding results in off-by-one error in accessing image elements. This is because the implementation(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L62-L66) computes two integers (representing the upper and lower bounds for interpolation) by ceiling and flooring a floating point value. For some values of `in`, `interpolation->upper[i]` might be smaller than `interpolation->lower[i]`. This is an issue if `interpolation->upper[i]` is capped at `in_size-1` as it means that `interpolation->lower[i]` points outside of the image. Then, in the interpolation code(https://github.com/tensorflow/tensorflow/blob/44b7f486c0143f68b56c34e2d01e146ee445134a/tensorflow/core/kernels/quantized_resize_bilinear_op.cc#L245-L264), this would result in heap buffer overflow. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29529>CVE-2021-29529</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-jfp7-4j67-8r3q</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file finalproject requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an end to end open source platform for machine learning an attacker can trigger a heap buffer overflow in tf raw ops quantizedresizebilinear by manipulating input values so that float rounding results in off by one error in accessing image elements this is because the implementation computes two integers representing the upper and lower bounds for interpolation by ceiling and flooring a floating point value for some values of in interpolation upper might be smaller than interpolation lower this is an issue if interpolation upper is capped at in size as it means that interpolation lower points outside of the image then in the interpolation code this would result in heap buffer overflow the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
110,025
23,854,240,817
IssuesEvent
2022-09-06 21:08:48
WordPress/openverse-catalog
https://api.github.com/repos/WordPress/openverse-catalog
opened
Consider setting category to 'illustration' for all svgs
🟩 priority: low ✨ goal: improvement 💻 aspect: code
## Current Situation <!-- Describe the part of the code you think should improve --> Related #614 Wikimedia sets the `category` for all records of filetype `svg` to "illustration". We should consider whether it's acceptable to do this for *all* providers, and if so we can move that logic to the `ImageStore` class. ## Suggested Improvement <!-- Describe your proposed change --> If we do this, then we can also remove Wikimedia's [extract_filetype](https://github.com/WordPress/openverse-catalog/blob/18decf9b70f409bc83809d14b7be3b65cb91d409/openverse_catalog/dags/providers/provider_api_scripts/wikimedia_commons.py#L349) method since the MediaStore will already handle extracting filetype from the media URL if `filetype` is None. ## Benefit <!-- Describe the benefit of the change (E.g., increase test coverage, reduce running time, etc.) --> Improved category population for more provider scripts, and allows removal of ad-hoc (and duplicated) code from Wikimedia. ## Implementation <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in implementing this feature.
1.0
Consider setting category to 'illustration' for all svgs - ## Current Situation <!-- Describe the part of the code you think should improve --> Related #614 Wikimedia sets the `category` for all records of filetype `svg` to "illustration". We should consider whether it's acceptable to do this for *all* providers, and if so we can move that logic to the `ImageStore` class. ## Suggested Improvement <!-- Describe your proposed change --> If we do this, then we can also remove Wikimedia's [extract_filetype](https://github.com/WordPress/openverse-catalog/blob/18decf9b70f409bc83809d14b7be3b65cb91d409/openverse_catalog/dags/providers/provider_api_scripts/wikimedia_commons.py#L349) method since the MediaStore will already handle extracting filetype from the media URL if `filetype` is None. ## Benefit <!-- Describe the benefit of the change (E.g., increase test coverage, reduce running time, etc.) --> Improved category population for more provider scripts, and allows removal of ad-hoc (and duplicated) code from Wikimedia. ## Implementation <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in implementing this feature.
non_process
consider setting category to illustration for all svgs current situation related wikimedia sets the category for all records of filetype svg to illustration we should consider whether it s acceptable to do this for all providers and if so we can move that logic to the imagestore class suggested improvement if we do this then we can also remove wikimedia s method since the mediastore will already handle extracting filetype from the media url if filetype is none benefit improved category population for more provider scripts and allows removal of ad hoc and duplicated code from wikimedia implementation 🙋 i would be interested in implementing this feature
0
253,725
27,300,814,243
IssuesEvent
2023-02-24 01:40:18
panasalap/linux-4.19.72_1
https://api.github.com/repos/panasalap/linux-4.19.72_1
closed
WS-2021-0462 (Medium) detected in linux-yoctov5.4.51 - autoclosed
security vulnerability
## WS-2021-0462 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Linux/Kernel is vulnerable to check ctr->cnr to avoid array index out of bound in drivers/isdn/capi/kcapi.c <p>Publish Date: 2021-11-29 <p>URL: <a href=https://github.com/gregkh/linux/commit/1f3e2e97c003f80c4b087092b225c8787ff91e4d>WS-2021-0462</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1002166">https://osv.dev/vulnerability/UVI-2021-1002166</a></p> <p>Release Date: 2021-11-29</p> <p>Fix Resolution: Linux/Kernel - v4.4.290, v4.9.288, v4.14.253, v4.19.214, v5.4.156, v5.10.76, v5.14.15, v5.15-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0462 (Medium) detected in linux-yoctov5.4.51 - autoclosed - ## WS-2021-0462 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/isdn/capi/kcapi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Linux/Kernel is vulnerable to check ctr->cnr to avoid array index out of bound in drivers/isdn/capi/kcapi.c <p>Publish Date: 2021-11-29 <p>URL: <a href=https://github.com/gregkh/linux/commit/1f3e2e97c003f80c4b087092b225c8787ff91e4d>WS-2021-0462</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1002166">https://osv.dev/vulnerability/UVI-2021-1002166</a></p> <p>Release Date: 2021-11-29</p> <p>Fix Resolution: Linux/Kernel - v4.4.290, v4.9.288, v4.14.253, v4.19.214, v5.4.156, v5.10.76, v5.14.15, v5.15-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in linux autoclosed ws medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers isdn capi kcapi c drivers isdn capi kcapi c vulnerability details in linux kernel is vulnerable to check ctr cnr to avoid array index out of bound in drivers isdn capi kcapi c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux kernel step up your open source security game with mend
0
121,237
25,941,646,322
IssuesEvent
2022-12-16 19:02:51
xygeni/xygeni-goat
https://api.github.com/repos/xygeni/xygeni-goat
opened
Add initial curated list of bad code
bad code
The goal is to add sources with intentionally flawed code, configurations, dependencies graph, IaC templates, etc. for learning about Software Supply Chain Security, and for running security tools. ### Location of files The initial set of bad code will be added under source/KIND, with KIND in { `secrets`, `misconfigs`, `bad_deps`, `iac_flaws` or `code_tamper` }. Under each directory, arbitrary directories could be defined for holding the bad code (for example each containing the code for a particular kind of defect). Please note that directories may need to be added so the files are properly scanned, so for example GitHub workflows are placed under `.github/workflows`, dependency descripts like package.json or package-lock.json need to have the exact name for the scanner to recognize them, etc. Examples: - `source/secrets/aws/aws_key.bad.json` (for this case, probably there is no need to add an `aws_key.bad.json`) - `source/bad_deps/scripts/bad/yarn.lock` (and corresponding package.json) and `source/bad_deps/scripts/good/yarn.lock` (and corresponding package.json) - note that bad and good directories instead of filename suffixes are used here, to preserve the file names. - `source/minconfigs/scriptinjection/.github/workflows/inline-script-injection.bad.yml` and corresponding `inline-script-injection.good.yml`. Note that .github/workflows should be used so the GitHub workflow YAML files are recognized. - `source/iac/terraform/aws/alb_protected_by_waf/bad/main.tf` (and others in same directory as needed for a coherent Terraform configuration) vs. `source/iac/terraform/aws/alb_protected_by_waf/good/main.tf`. All files belonging to the same configuration must be placed under the same directory. ### Expections file - expected.yaml An `expected.yaml` file could be added for each issue example, with two lists: `fail` with the items that have security issues, `pass` for the examples with fixed code. Each item listed could be the resource ID in IaC files, a path in the hierarchy of a configuration file like a JSON or YAML, line number or lines range. Anything helping the user to locate where is the problem, even to use the `expected.yaml` to automate verification of results from a scanner.
1.0
Add initial curated list of bad code - The goal is to add sources with intentionally flawed code, configurations, dependencies graph, IaC templates, etc. for learning about Software Supply Chain Security, and for running security tools. ### Location of files The initial set of bad code will be added under source/KIND, with KIND in { `secrets`, `misconfigs`, `bad_deps`, `iac_flaws` or `code_tamper` }. Under each directory, arbitrary directories could be defined for holding the bad code (for example each containing the code for a particular kind of defect). Please note that directories may need to be added so the files are properly scanned, so for example GitHub workflows are placed under `.github/workflows`, dependency descripts like package.json or package-lock.json need to have the exact name for the scanner to recognize them, etc. Examples: - `source/secrets/aws/aws_key.bad.json` (for this case, probably there is no need to add an `aws_key.bad.json`) - `source/bad_deps/scripts/bad/yarn.lock` (and corresponding package.json) and `source/bad_deps/scripts/good/yarn.lock` (and corresponding package.json) - note that bad and good directories instead of filename suffixes are used here, to preserve the file names. - `source/minconfigs/scriptinjection/.github/workflows/inline-script-injection.bad.yml` and corresponding `inline-script-injection.good.yml`. Note that .github/workflows should be used so the GitHub workflow YAML files are recognized. - `source/iac/terraform/aws/alb_protected_by_waf/bad/main.tf` (and others in same directory as needed for a coherent Terraform configuration) vs. `source/iac/terraform/aws/alb_protected_by_waf/good/main.tf`. All files belonging to the same configuration must be placed under the same directory. ### Expections file - expected.yaml An `expected.yaml` file could be added for each issue example, with two lists: `fail` with the items that have security issues, `pass` for the examples with fixed code. Each item listed could be the resource ID in IaC files, a path in the hierarchy of a configuration file like a JSON or YAML, line number or lines range. Anything helping the user to locate where is the problem, even to use the `expected.yaml` to automate verification of results from a scanner.
non_process
add initial curated list of bad code the goal is to add sources with intentionally flawed code configurations dependencies graph iac templates etc for learning about software supply chain security and for running security tools location of files the initial set of bad code will be added under source kind with kind in secrets misconfigs bad deps iac flaws or code tamper under each directory arbitrary directories could be defined for holding the bad code for example each containing the code for a particular kind of defect please note that directories may need to be added so the files are properly scanned so for example github workflows are placed under github workflows dependency descripts like package json or package lock json need to have the exact name for the scanner to recognize them etc examples source secrets aws aws key bad json for this case probably there is no need to add an aws key bad json source bad deps scripts bad yarn lock and corresponding package json and source bad deps scripts good yarn lock and corresponding package json note that bad and good directories instead of filename suffixes are used here to preserve the file names source minconfigs scriptinjection github workflows inline script injection bad yml and corresponding inline script injection good yml note that github workflows should be used so the github workflow yaml files are recognized source iac terraform aws alb protected by waf bad main tf and others in same directory as needed for a coherent terraform configuration vs source iac terraform aws alb protected by waf good main tf all files belonging to the same configuration must be placed under the same directory expections file expected yaml an expected yaml file could be added for each issue example with two lists fail with the items that have security issues pass for the examples with fixed code each item listed could be the resource id in iac files a path in the hierarchy of a configuration file like a json or yaml line number or lines range anything helping the user to locate where is the problem even to use the expected yaml to automate verification of results from a scanner
0
17,973
23,984,561,074
IssuesEvent
2022-09-13 17:51:54
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
opened
Configure blunderbuss and CODEOWNERS on this repository
type: process
This repo will contain hundreds of client libraries and will want some help assigning issues and PRs to the correct reviewers.
1.0
Configure blunderbuss and CODEOWNERS on this repository - This repo will contain hundreds of client libraries and will want some help assigning issues and PRs to the correct reviewers.
process
configure blunderbuss and codeowners on this repository this repo will contain hundreds of client libraries and will want some help assigning issues and prs to the correct reviewers
1
18,770
24,674,394,274
IssuesEvent
2022-10-18 15:51:04
deepset-ai/haystack
https://api.github.com/repos/deepset-ai/haystack
closed
Page number in document meta data not correct
type:bug topic:file_converter topic:preprocessing
**Describe the bug** The new "add page number" features, implemented in #2932, seem to have a bug in combination with the `PDFToTextConverter` **Error message** Wrong page number in metadata **To Reproduce** Use the following haystack components: ```python elastic_search_document_store = ElasticsearchDocumentStore(...) retriever = BM25Retriever(document_store=elastic_search_document_store, all_terms_must_match=False) converter = PDFToTextConverter(remove_numeric_tables=True, valid_languages=["de", "en"]) pre_processor = PreProcessor(split_length=100, add_page_number=True) # here we use the new page number flag reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2", use_gpu=False) ``` Now we upload, convert and pre-process the document [tesla_annual_report.pdf](https://github.com/deepset-ai/haystack/files/9430730/tesla_annual_report.pdf) ```python docs: list[Document] = converter.convert(Path(tmp_file.name), meta={'filename': filename}) docs_pre_processed: list[Document] = pre_processor.process(docs) elastic_search_document_store.write_documents(docs_pre_processed) ``` Now we search for the search string *'financial statement'*: ```python search_query = "financial statement" result_documents: list[Document] = retriever.retrieve(query=search_query, top_k=30) answers: list[Answer] = self.reader.predict(search_query, result_documents)['answers'] ``` if we have a look at the top answer (and the corresponding `Document`) we see the answer ``` 'reliability of financial reporting and the preparation of financial statements for external purposes in accordance with generally accepted accounting principles' ``` and the meta data **page** in the `Document` is **45**. If we though look into the pdf-document and search manually for the answer, we find the answer on page **48**. The bug appears frequently, with a number of other pdf. **FAQ Check** - [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)? **System:** - OS: *Ubuntu 20.04* - GPU/CPU: *both* - Haystack version (commit or version number): *master, c0fbe45c029247428cee3823292bd574b1846c38* - DocumentStore: *see "To Reproduce"* - Reader: *see "To Reproduce"* - Retriever: *see "To Reproduce"*
1.0
Page number in document meta data not correct - **Describe the bug** The new "add page number" features, implemented in #2932, seem to have a bug in combination with the `PDFToTextConverter` **Error message** Wrong page number in metadata **To Reproduce** Use the following haystack components: ```python elastic_search_document_store = ElasticsearchDocumentStore(...) retriever = BM25Retriever(document_store=elastic_search_document_store, all_terms_must_match=False) converter = PDFToTextConverter(remove_numeric_tables=True, valid_languages=["de", "en"]) pre_processor = PreProcessor(split_length=100, add_page_number=True) # here we use the new page number flag reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2", use_gpu=False) ``` Now we upload, convert and pre-process the document [tesla_annual_report.pdf](https://github.com/deepset-ai/haystack/files/9430730/tesla_annual_report.pdf) ```python docs: list[Document] = converter.convert(Path(tmp_file.name), meta={'filename': filename}) docs_pre_processed: list[Document] = pre_processor.process(docs) elastic_search_document_store.write_documents(docs_pre_processed) ``` Now we search for the search string *'financial statement'*: ```python search_query = "financial statement" result_documents: list[Document] = retriever.retrieve(query=search_query, top_k=30) answers: list[Answer] = self.reader.predict(search_query, result_documents)['answers'] ``` if we have a look at the top answer (and the corresponding `Document`) we see the answer ``` 'reliability of financial reporting and the preparation of financial statements for external purposes in accordance with generally accepted accounting principles' ``` and the meta data **page** in the `Document` is **45**. If we though look into the pdf-document and search manually for the answer, we find the answer on page **48**. The bug appears frequently, with a number of other pdf. **FAQ Check** - [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)? **System:** - OS: *Ubuntu 20.04* - GPU/CPU: *both* - Haystack version (commit or version number): *master, c0fbe45c029247428cee3823292bd574b1846c38* - DocumentStore: *see "To Reproduce"* - Reader: *see "To Reproduce"* - Retriever: *see "To Reproduce"*
process
page number in document meta data not correct describe the bug the new add page number features implemented in seem to have a bug in combination with the pdftotextconverter error message wrong page number in metadata to reproduce use the following haystack components python elastic search document store elasticsearchdocumentstore retriever document store elastic search document store all terms must match false converter pdftotextconverter remove numeric tables true valid languages pre processor preprocessor split length add page number true here we use the new page number flag reader farmreader model name or path deepset roberta base use gpu false now we upload convert and pre process the document python docs list converter convert path tmp file name meta filename filename docs pre processed list pre processor process docs elastic search document store write documents docs pre processed now we search for the search string financial statement python search query financial statement result documents list retriever retrieve query search query top k answers list self reader predict search query result documents if we have a look at the top answer and the corresponding document we see the answer reliability of financial reporting and the preparation of financial statements for external purposes in accordance with generally accepted accounting principles and the meta data page in the document is if we though look into the pdf document and search manually for the answer we find the answer on page the bug appears frequently with a number of other pdf faq check have you had a look at system os ubuntu gpu cpu both haystack version commit or version number master documentstore see to reproduce reader see to reproduce retriever see to reproduce
1
19,079
25,119,464,349
IssuesEvent
2022-11-09 06:39:39
streamnative/flink
https://api.github.com/repos/streamnative/flink
opened
[Connector] Do not fail job when PulsarAdmin API fails when getting the data.
compute/data-processing
Currently, Pulsar connector periodically queries the admin api to detect partition changes. However, this can cause the pipeline to fail if the admin api fail. @weixiangchen Reported a admin 502 when running pipeline for longer peirods (30 minutes). We should discuss whether we want to 1) fail the job 2) restart the job 3) report the exception and go on when similiar 502 exception happens.
1.0
[Connector] Do not fail job when PulsarAdmin API fails when getting the data. - Currently, Pulsar connector periodically queries the admin api to detect partition changes. However, this can cause the pipeline to fail if the admin api fail. @weixiangchen Reported a admin 502 when running pipeline for longer peirods (30 minutes). We should discuss whether we want to 1) fail the job 2) restart the job 3) report the exception and go on when similiar 502 exception happens.
process
do not fail job when pulsaradmin api fails when getting the data currently pulsar connector periodically queries the admin api to detect partition changes however this can cause the pipeline to fail if the admin api fail weixiangchen reported a admin when running pipeline for longer peirods minutes we should discuss whether we want to fail the job restart the job report the exception and go on when similiar exception happens
1
12,914
15,287,552,000
IssuesEvent
2021-02-23 15:53:25
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
closed
vdc user to be able to set autoscale limit
process_wontfix type_feature
the user may need to specify an autoscale limit so they don't grow indefinitely
1.0
vdc user to be able to set autoscale limit - the user may need to specify an autoscale limit so they don't grow indefinitely
process
vdc user to be able to set autoscale limit the user may need to specify an autoscale limit so they don t grow indefinitely
1
107,836
9,231,248,308
IssuesEvent
2019-03-13 01:29:26
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
ICE Type parameter out of range when substituting
A-typesystem E-needstest I-ICE T-compiler
Both nightly and stable ICE on this code: ```rust use once_cell::sync::OnceCell; use std::collections::HashMap; type Cache<K, T> = OnceCell<HashMap<K, T>>; trait Provider<K, T> { fn new_cache() -> Cache<K, T> { OnceCell::INIT } } struct Fib; impl Provider<u32, u32> for Fib {} fn main() { let _ = Fib::new_cache(); } ``` I expected to see this happen: This should either compile or throw an Error explaining why it is invalid. Instead, this happened: Both `cargo build` and `cargo +nightly build` throw an ICE. ## Meta installed dependencies: ```toml once_cell = "0.1.7" ``` `rustc --version --verbose`: rustc 1.32.0 (9fda7c223 2019-01-16) binary: rustc commit-hash: 9fda7c2237db910e41d6a712e9a2139b352e558b commit-date: 2019-01-16 host: x86_64-unknown-linux-gnu release: 1.32.0 LLVM version: 8.0 `rustc +nightly --version --verbose`: rustc 1.33.0-nightly (0c0c58528 2019-01-19) binary: rustc commit-hash: 0c0c585281740aba4c91cbfd385f5e1fbe313d11 commit-date: 2019-01-19 host: x86_64-unknown-linux-gnu release: 1.33.0-nightly LLVM version: 8.0 Backtrace: ``` error: internal compiler error: src/librustc/ty/subst.rs:491: Type parameter `K/#1` (K/1) out of range when substituting (root type=Some(std::option::Option<std::collections::HashMap<K, T>>)) substs=[std::collections::HashMap<K, T>] thread 'main' panicked at 'Box<Any>', src/librustc_errors/lib.rs:538:9 stack backtrace: 0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace at src/libstd/sys/unix/backtrace/tracing/gcc_s.rs:49 1: std::sys_common::backtrace::_print at src/libstd/sys_common/backtrace.rs:71 2: std::panicking::default_hook::{{closure}} at src/libstd/sys_common/backtrace.rs:59 at src/libstd/panicking.rs:211 3: std::panicking::default_hook at src/libstd/panicking.rs:227 4: rustc::util::common::panic_hook 5: std::panicking::rust_panic_with_hook at src/libstd/panicking.rs:495 6: std::panicking::begin_panic 7: rustc_errors::Handler::span_bug 8: rustc::util::bug::opt_span_bug_fmt::{{closure}} 9: rustc::ty::context::tls::with_opt::{{closure}} 10: rustc::ty::context::tls::with_context_opt 11: rustc::ty::context::tls::with_opt 12: rustc::util::bug::opt_span_bug_fmt 13: rustc::util::bug::span_bug_fmt 14: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty 15: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter 16: rustc::ty::fold::TypeFoldable::fold_with 17: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty 18: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter 19: rustc::ty::fold::TypeFoldable::fold_with 20: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty 21: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter 22: rustc::ty::fold::TypeFoldable::fold_with 23: rustc::traits::codegen::<impl rustc::ty::context::TyCtxt<'a, 'tcx, 'tcx>>::subst_and_normalize_erasing_regions 24: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::resolve 25: rustc_mir::interpret::step::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::run 26: rustc_mir::const_eval::eval_body_using_ecx 27: rustc_mir::const_eval::const_eval_raw_provider 28: rustc::ty::query::__query_compute::const_eval_raw 29: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::const_eval_raw<'tcx>>::compute 30: rustc::ty::context::tls::with_context 31: rustc::dep_graph::graph::DepGraph::with_task_impl 32: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start 33: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job 34: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query 35: rustc::ty::query::TyCtxtAt::const_eval_raw 36: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_eval_raw 37: rustc_mir::interpret::operand::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_value_to_op 38: rustc_mir::const_eval::const_to_op 39: rustc_mir::transform::const_prop::ConstPropagator::eval_constant 40: <rustc_mir::transform::const_prop::ConstPropagator<'b, 'a, 'tcx> as rustc::mir::visit::Visitor<'tcx>>::visit_statement 41: <rustc_mir::transform::const_prop::ConstProp as rustc_mir::transform::MirPass>::run_pass 42: rustc_mir::transform::run_passes::{{closure}} 43: rustc_mir::transform::run_passes 44: rustc_mir::transform::optimized_mir 45: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::optimized_mir<'tcx>>::compute 46: rustc::ty::context::tls::with_context 47: rustc::dep_graph::graph::DepGraph::with_task_impl 48: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start 49: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job 50: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::try_get_query 51: rustc::ty::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::instance_mir 52: rustc_mir::monomorphize::collector::collect_items_rec 53: rustc_mir::monomorphize::collector::collect_items_rec 54: rustc_mir::monomorphize::collector::collect_crate_mono_items::{{closure}} 55: rustc::util::common::time 56: rustc_mir::monomorphize::collector::collect_crate_mono_items 57: rustc::util::common::time 58: rustc_mir::monomorphize::partitioning::collect_and_partition_mono_items 59: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::collect_and_partition_mono_items<'tcx>>::compute 60: rustc::ty::context::tls::with_context 61: rustc::dep_graph::graph::DepGraph::with_task_impl 62: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start 63: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job 64: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query 65: rustc_codegen_ssa::base::codegen_crate 66: <rustc_codegen_llvm::LlvmCodegenBackend as rustc_codegen_utils::codegen_backend::CodegenBackend>::codegen_crate 67: rustc::util::common::time 68: rustc_driver::driver::phase_4_codegen 69: rustc_driver::driver::compile_input::{{closure}} 70: rustc::ty::context::tls::enter_context 71: <std::thread::local::LocalKey<T>>::with 72: rustc::ty::context::TyCtxt::create_and_enter 73: rustc_driver::driver::compile_input 74: rustc_driver::run_compiler_with_pool 75: <scoped_tls::ScopedKey<T>>::set 76: rustc_driver::run_compiler 77: rustc_driver::monitor::{{closure}} 78: __rust_maybe_catch_panic at src/libpanic_unwind/lib.rs:102 79: rustc_driver::run 80: rustc_driver::main 81: std::rt::lang_start::{{closure}} 82: std::panicking::try::do_call at src/libstd/rt.rs:59 at src/libstd/panicking.rs:310 83: __rust_maybe_catch_panic at src/libpanic_unwind/lib.rs:102 84: std::rt::lang_start_internal at src/libstd/panicking.rs:289 at src/libstd/panic.rs:398 at src/libstd/rt.rs:58 85: main 86: __libc_start_main 87: <unknown> query stack during panic: #0 [const_eval_raw] const-evaluating `<once_cell::imp::OnceCell<T>>::INIT` --> src/main.rs:8:9 | 8 | OnceCell::INIT | ^^^^^^^^^^^^^^ #1 [optimized_mir] processing `Provider::new_cache` #2 [collect_and_partition_mono_items] collect_and_partition_mono_items end of query stack ``` Note: There are multiple issues with similar titles, sorry if this is a duplicate. Most other issues I could find seem to only affect nightly, this crashes on stable, too. #57700 panics during `mir_borrowck`, this in `optimized_mir`.
1.0
ICE Type parameter out of range when substituting - Both nightly and stable ICE on this code: ```rust use once_cell::sync::OnceCell; use std::collections::HashMap; type Cache<K, T> = OnceCell<HashMap<K, T>>; trait Provider<K, T> { fn new_cache() -> Cache<K, T> { OnceCell::INIT } } struct Fib; impl Provider<u32, u32> for Fib {} fn main() { let _ = Fib::new_cache(); } ``` I expected to see this happen: This should either compile or throw an Error explaining why it is invalid. Instead, this happened: Both `cargo build` and `cargo +nightly build` throw an ICE. ## Meta installed dependencies: ```toml once_cell = "0.1.7" ``` `rustc --version --verbose`: rustc 1.32.0 (9fda7c223 2019-01-16) binary: rustc commit-hash: 9fda7c2237db910e41d6a712e9a2139b352e558b commit-date: 2019-01-16 host: x86_64-unknown-linux-gnu release: 1.32.0 LLVM version: 8.0 `rustc +nightly --version --verbose`: rustc 1.33.0-nightly (0c0c58528 2019-01-19) binary: rustc commit-hash: 0c0c585281740aba4c91cbfd385f5e1fbe313d11 commit-date: 2019-01-19 host: x86_64-unknown-linux-gnu release: 1.33.0-nightly LLVM version: 8.0 Backtrace: ``` error: internal compiler error: src/librustc/ty/subst.rs:491: Type parameter `K/#1` (K/1) out of range when substituting (root type=Some(std::option::Option<std::collections::HashMap<K, T>>)) substs=[std::collections::HashMap<K, T>] thread 'main' panicked at 'Box<Any>', src/librustc_errors/lib.rs:538:9 stack backtrace: 0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace at src/libstd/sys/unix/backtrace/tracing/gcc_s.rs:49 1: std::sys_common::backtrace::_print at src/libstd/sys_common/backtrace.rs:71 2: std::panicking::default_hook::{{closure}} at src/libstd/sys_common/backtrace.rs:59 at src/libstd/panicking.rs:211 3: std::panicking::default_hook at src/libstd/panicking.rs:227 4: rustc::util::common::panic_hook 5: std::panicking::rust_panic_with_hook at src/libstd/panicking.rs:495 6: std::panicking::begin_panic 7: rustc_errors::Handler::span_bug 8: rustc::util::bug::opt_span_bug_fmt::{{closure}} 9: rustc::ty::context::tls::with_opt::{{closure}} 10: rustc::ty::context::tls::with_context_opt 11: rustc::ty::context::tls::with_opt 12: rustc::util::bug::opt_span_bug_fmt 13: rustc::util::bug::span_bug_fmt 14: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty 15: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter 16: rustc::ty::fold::TypeFoldable::fold_with 17: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty 18: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter 19: rustc::ty::fold::TypeFoldable::fold_with 20: <rustc::ty::subst::SubstFolder<'a, 'gcx, 'tcx> as rustc::ty::fold::TypeFolder<'gcx, 'tcx>>::fold_ty 21: <smallvec::SmallVec<A> as core::iter::traits::FromIterator<<A as smallvec::Array>::Item>>::from_iter 22: rustc::ty::fold::TypeFoldable::fold_with 23: rustc::traits::codegen::<impl rustc::ty::context::TyCtxt<'a, 'tcx, 'tcx>>::subst_and_normalize_erasing_regions 24: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::resolve 25: rustc_mir::interpret::step::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::run 26: rustc_mir::const_eval::eval_body_using_ecx 27: rustc_mir::const_eval::const_eval_raw_provider 28: rustc::ty::query::__query_compute::const_eval_raw 29: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::const_eval_raw<'tcx>>::compute 30: rustc::ty::context::tls::with_context 31: rustc::dep_graph::graph::DepGraph::with_task_impl 32: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start 33: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job 34: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query 35: rustc::ty::query::TyCtxtAt::const_eval_raw 36: <rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_eval_raw 37: rustc_mir::interpret::operand::<impl rustc_mir::interpret::eval_context::EvalContext<'a, 'mir, 'tcx, M>>::const_value_to_op 38: rustc_mir::const_eval::const_to_op 39: rustc_mir::transform::const_prop::ConstPropagator::eval_constant 40: <rustc_mir::transform::const_prop::ConstPropagator<'b, 'a, 'tcx> as rustc::mir::visit::Visitor<'tcx>>::visit_statement 41: <rustc_mir::transform::const_prop::ConstProp as rustc_mir::transform::MirPass>::run_pass 42: rustc_mir::transform::run_passes::{{closure}} 43: rustc_mir::transform::run_passes 44: rustc_mir::transform::optimized_mir 45: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::optimized_mir<'tcx>>::compute 46: rustc::ty::context::tls::with_context 47: rustc::dep_graph::graph::DepGraph::with_task_impl 48: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start 49: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job 50: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::try_get_query 51: rustc::ty::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::instance_mir 52: rustc_mir::monomorphize::collector::collect_items_rec 53: rustc_mir::monomorphize::collector::collect_items_rec 54: rustc_mir::monomorphize::collector::collect_crate_mono_items::{{closure}} 55: rustc::util::common::time 56: rustc_mir::monomorphize::collector::collect_crate_mono_items 57: rustc::util::common::time 58: rustc_mir::monomorphize::partitioning::collect_and_partition_mono_items 59: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors<'tcx> for rustc::ty::query::queries::collect_and_partition_mono_items<'tcx>>::compute 60: rustc::ty::context::tls::with_context 61: rustc::dep_graph::graph::DepGraph::with_task_impl 62: <rustc::ty::query::plumbing::JobOwner<'a, 'tcx, Q>>::start 63: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::force_query_with_job 64: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt<'a, 'gcx, 'tcx>>::get_query 65: rustc_codegen_ssa::base::codegen_crate 66: <rustc_codegen_llvm::LlvmCodegenBackend as rustc_codegen_utils::codegen_backend::CodegenBackend>::codegen_crate 67: rustc::util::common::time 68: rustc_driver::driver::phase_4_codegen 69: rustc_driver::driver::compile_input::{{closure}} 70: rustc::ty::context::tls::enter_context 71: <std::thread::local::LocalKey<T>>::with 72: rustc::ty::context::TyCtxt::create_and_enter 73: rustc_driver::driver::compile_input 74: rustc_driver::run_compiler_with_pool 75: <scoped_tls::ScopedKey<T>>::set 76: rustc_driver::run_compiler 77: rustc_driver::monitor::{{closure}} 78: __rust_maybe_catch_panic at src/libpanic_unwind/lib.rs:102 79: rustc_driver::run 80: rustc_driver::main 81: std::rt::lang_start::{{closure}} 82: std::panicking::try::do_call at src/libstd/rt.rs:59 at src/libstd/panicking.rs:310 83: __rust_maybe_catch_panic at src/libpanic_unwind/lib.rs:102 84: std::rt::lang_start_internal at src/libstd/panicking.rs:289 at src/libstd/panic.rs:398 at src/libstd/rt.rs:58 85: main 86: __libc_start_main 87: <unknown> query stack during panic: #0 [const_eval_raw] const-evaluating `<once_cell::imp::OnceCell<T>>::INIT` --> src/main.rs:8:9 | 8 | OnceCell::INIT | ^^^^^^^^^^^^^^ #1 [optimized_mir] processing `Provider::new_cache` #2 [collect_and_partition_mono_items] collect_and_partition_mono_items end of query stack ``` Note: There are multiple issues with similar titles, sorry if this is a duplicate. Most other issues I could find seem to only affect nightly, this crashes on stable, too. #57700 panics during `mir_borrowck`, this in `optimized_mir`.
non_process
ice type parameter out of range when substituting both nightly and stable ice on this code rust use once cell sync oncecell use std collections hashmap type cache oncecell trait provider fn new cache cache oncecell init struct fib impl provider for fib fn main let fib new cache i expected to see this happen this should either compile or throw an error explaining why it is invalid instead this happened both cargo build and cargo nightly build throw an ice meta installed dependencies toml once cell rustc version verbose rustc binary rustc commit hash commit date host unknown linux gnu release llvm version rustc nightly version verbose rustc nightly binary rustc commit hash commit date host unknown linux gnu release nightly llvm version backtrace error internal compiler error src librustc ty subst rs type parameter k k out of range when substituting root type some std option option substs thread main panicked at box src librustc errors lib rs stack backtrace std sys unix backtrace tracing imp unwind backtrace at src libstd sys unix backtrace tracing gcc s rs std sys common backtrace print at src libstd sys common backtrace rs std panicking default hook closure at src libstd sys common backtrace rs at src libstd panicking rs std panicking default hook at src libstd panicking rs rustc util common panic hook std panicking rust panic with hook at src libstd panicking rs std panicking begin panic rustc errors handler span bug rustc util bug opt span bug fmt closure rustc ty context tls with opt closure rustc ty context tls with context opt rustc ty context tls with opt rustc util bug opt span bug fmt rustc util bug span bug fmt as rustc ty fold typefolder fold ty as core iter traits fromiterator item from iter rustc ty fold typefoldable fold with as rustc ty fold typefolder fold ty as core iter traits fromiterator item from iter rustc ty fold typefoldable fold with as rustc ty fold typefolder fold ty as core iter traits fromiterator item from iter rustc ty fold typefoldable fold with rustc traits codegen subst and normalize erasing regions resolve rustc mir interpret step run rustc mir const eval eval body using ecx rustc mir const eval const eval raw provider rustc ty query query compute const eval raw rustc ty query for rustc ty query queries const eval raw compute rustc ty context tls with context rustc dep graph graph depgraph with task impl start rustc ty query plumbing force query with job rustc ty query plumbing get query rustc ty query tyctxtat const eval raw const eval raw rustc mir interpret operand const value to op rustc mir const eval const to op rustc mir transform const prop constpropagator eval constant as rustc mir visit visitor visit statement run pass rustc mir transform run passes closure rustc mir transform run passes rustc mir transform optimized mir rustc ty query for rustc ty query queries optimized mir compute rustc ty context tls with context rustc dep graph graph depgraph with task impl start rustc ty query plumbing force query with job rustc ty query plumbing try get query rustc ty instance mir rustc mir monomorphize collector collect items rec rustc mir monomorphize collector collect items rec rustc mir monomorphize collector collect crate mono items closure rustc util common time rustc mir monomorphize collector collect crate mono items rustc util common time rustc mir monomorphize partitioning collect and partition mono items rustc ty query for rustc ty query queries collect and partition mono items compute rustc ty context tls with context rustc dep graph graph depgraph with task impl start rustc ty query plumbing force query with job rustc ty query plumbing get query rustc codegen ssa base codegen crate codegen crate rustc util common time rustc driver driver phase codegen rustc driver driver compile input closure rustc ty context tls enter context with rustc ty context tyctxt create and enter rustc driver driver compile input rustc driver run compiler with pool set rustc driver run compiler rustc driver monitor closure rust maybe catch panic at src libpanic unwind lib rs rustc driver run rustc driver main std rt lang start closure std panicking try do call at src libstd rt rs at src libstd panicking rs rust maybe catch panic at src libpanic unwind lib rs std rt lang start internal at src libstd panicking rs at src libstd panic rs at src libstd rt rs main libc start main query stack during panic const evaluating init src main rs oncecell init processing provider new cache collect and partition mono items end of query stack note there are multiple issues with similar titles sorry if this is a duplicate most other issues i could find seem to only affect nightly this crashes on stable too panics during mir borrowck this in optimized mir
0
621,590
19,592,255,948
IssuesEvent
2022-01-05 14:13:53
WordPress/openverse-frontend
https://api.github.com/repos/WordPress/openverse-frontend
closed
Implement redesign and new components
🟧 priority: high ✨ goal: improvement 🕹 aspect: interface
This is a meta ticket to track the ongoing redesign efforts in Openverse. This ticket is very much a work in progress and is currently being used to track missing, ongoing issues with the redesign. ## Design Todos ### General - [ ] #256 - [x] #273 - [ ] #349 - [ ] Simplified homepage (no featured collections yet) - [ ] Redesigned content pages (anything text-heavy like `/sources` and so on) - These shouldn't need much, just a refresh with the latest colors, sizing, and so on ### Mature content / copyright claims - [ ] Add "report this {image|audio}" flow to the single result pages <img width="372" alt="CleanShot 2021-10-29 at 10 34 29@2x" src="https://user-images.githubusercontent.com/6351754/139453661-f4270159-9a87-4b87-a82f-f8371e4e0d69.png"> - [ ] Add back filter to enable mature content (this can go in the filter sidebar) - [ ] The reporting modal (needs to be updated in the new popover style <img width="385" alt="CleanShot 2021-10-29 at 10 31 05@2x" src="https://user-images.githubusercontent.com/6351754/139453202-b603c027-070f-4560-a843-52b3d0541304.png">)
1.0
Implement redesign and new components - This is a meta ticket to track the ongoing redesign efforts in Openverse. This ticket is very much a work in progress and is currently being used to track missing, ongoing issues with the redesign. ## Design Todos ### General - [ ] #256 - [x] #273 - [ ] #349 - [ ] Simplified homepage (no featured collections yet) - [ ] Redesigned content pages (anything text-heavy like `/sources` and so on) - These shouldn't need much, just a refresh with the latest colors, sizing, and so on ### Mature content / copyright claims - [ ] Add "report this {image|audio}" flow to the single result pages <img width="372" alt="CleanShot 2021-10-29 at 10 34 29@2x" src="https://user-images.githubusercontent.com/6351754/139453661-f4270159-9a87-4b87-a82f-f8371e4e0d69.png"> - [ ] Add back filter to enable mature content (this can go in the filter sidebar) - [ ] The reporting modal (needs to be updated in the new popover style <img width="385" alt="CleanShot 2021-10-29 at 10 31 05@2x" src="https://user-images.githubusercontent.com/6351754/139453202-b603c027-070f-4560-a843-52b3d0541304.png">)
non_process
implement redesign and new components this is a meta ticket to track the ongoing redesign efforts in openverse this ticket is very much a work in progress and is currently being used to track missing ongoing issues with the redesign design todos general simplified homepage no featured collections yet redesigned content pages anything text heavy like sources and so on these shouldn t need much just a refresh with the latest colors sizing and so on mature content copyright claims add report this image audio flow to the single result pages img width alt cleanshot at src add back filter to enable mature content this can go in the filter sidebar the reporting modal needs to be updated in the new popover style img width alt cleanshot at src
0
11,553
14,435,280,100
IssuesEvent
2020-12-07 08:30:19
linuxdeepin/developer-center
https://api.github.com/repos/linuxdeepin/developer-center
closed
Control panel crash when click icons
bug | Ports bug | functional behavior other | delay processing suggest | functional behavior
Hi, i'm using deepin DE on Arch Linux, at the last upgrade when i switch from deepin-control-center-5.2.0.1 to deepin-control-center-5.2.0.3 i had a problem: the program crash when i select the account icon. When i downgrade to the older version it works, i think that is a problem of the newest version.
1.0
Control panel crash when click icons - Hi, i'm using deepin DE on Arch Linux, at the last upgrade when i switch from deepin-control-center-5.2.0.1 to deepin-control-center-5.2.0.3 i had a problem: the program crash when i select the account icon. When i downgrade to the older version it works, i think that is a problem of the newest version.
process
control panel crash when click icons hi i m using deepin de on arch linux at the last upgrade when i switch from deepin control center to deepin control center i had a problem the program crash when i select the account icon when i downgrade to the older version it works i think that is a problem of the newest version
1
23,709
12,086,734,667
IssuesEvent
2020-04-18 11:28:20
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Dart high memory when open VS Code (macos)
dependency: dart framework platform-mac severe: performance tool
Hi, I'm getting slow when I open vs code that has installed Flutter plugins. I look memory size of dart, consumed : 1,05 GB Details : ``` cwd / txt /Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/dart txt /Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/analysis_server.dart.snapshot txt /Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/kernel-service.dart.snapshot txt /usr/lib/dyld txt /private/var/db/dyld/dyld_shared_cache_x86_64 0 ->0x6615e580cac70bb5 1 ->0x6615e580cac71515 2 ->0x6615e580cac707cd 3 ->0x6615e580cac759a5 4 ->0x6615e580cac75a65 5 count=0, state=0xa 6 ->0x6615e580cac785e5 7 ->0x6615e580cac75b25 10 ->0x6615e580cac758e5 11 ->0x6615e580cac78525 12 ->0x6615e580cac786a5 13 ->0x6615e580cac78465 37 /Applications/Visual Studio Code.app/Contents/Resources/app/node_modules.asar ``` and here the picture. Please fix it, if there are some bugs. Thank you. ![Screen Shot 2019-10-21 at 09 26 07](https://user-images.githubusercontent.com/33727072/67172258-d7324600-f3e4-11e9-90da-fe8202b70713.png)
True
Dart high memory when open VS Code (macos) - Hi, I'm getting slow when I open vs code that has installed Flutter plugins. I look memory size of dart, consumed : 1,05 GB Details : ``` cwd / txt /Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/dart txt /Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/analysis_server.dart.snapshot txt /Users/macosx/flutter-1.10.4/bin/cache/dart-sdk/bin/snapshots/kernel-service.dart.snapshot txt /usr/lib/dyld txt /private/var/db/dyld/dyld_shared_cache_x86_64 0 ->0x6615e580cac70bb5 1 ->0x6615e580cac71515 2 ->0x6615e580cac707cd 3 ->0x6615e580cac759a5 4 ->0x6615e580cac75a65 5 count=0, state=0xa 6 ->0x6615e580cac785e5 7 ->0x6615e580cac75b25 10 ->0x6615e580cac758e5 11 ->0x6615e580cac78525 12 ->0x6615e580cac786a5 13 ->0x6615e580cac78465 37 /Applications/Visual Studio Code.app/Contents/Resources/app/node_modules.asar ``` and here the picture. Please fix it, if there are some bugs. Thank you. ![Screen Shot 2019-10-21 at 09 26 07](https://user-images.githubusercontent.com/33727072/67172258-d7324600-f3e4-11e9-90da-fe8202b70713.png)
non_process
dart high memory when open vs code macos hi i m getting slow when i open vs code that has installed flutter plugins i look memory size of dart consumed gb details cwd txt users macosx flutter bin cache dart sdk bin dart txt users macosx flutter bin cache dart sdk bin snapshots analysis server dart snapshot txt users macosx flutter bin cache dart sdk bin snapshots kernel service dart snapshot txt usr lib dyld txt private var db dyld dyld shared cache count state applications visual studio code app contents resources app node modules asar and here the picture please fix it if there are some bugs thank you
0
164,423
13,942,219,347
IssuesEvent
2020-10-22 20:40:04
TheRedDaemon/LittleCrusaderAsi
https://api.github.com/repos/TheRedDaemon/LittleCrusaderAsi
opened
Readme for implementation concepts still needs to be written.
documentation
It is already linkend and should contain something. Otherwise it is not helpful...
1.0
Readme for implementation concepts still needs to be written. - It is already linkend and should contain something. Otherwise it is not helpful...
non_process
readme for implementation concepts still needs to be written it is already linkend and should contain something otherwise it is not helpful
0
31,722
5,989,019,130
IssuesEvent
2017-06-02 07:18:11
icebob/vue-form-generator
https://api.github.com/repos/icebob/vue-form-generator
closed
update docs
difficulty: easy documentation
There are some recent changes what we need to update in documentation too. - [x] `change` property in schema of file input #173 - [x] supported async validators #171 - [x] custom validation message for fields #169 - [x] supported string-based validators #167 - [x] vue-multiselect fixed #30 but in beta15, but they expanded [CSS from JS bundle](http://monterail.github.io/vue-multiselect/#sub-getting-started).
1.0
update docs - There are some recent changes what we need to update in documentation too. - [x] `change` property in schema of file input #173 - [x] supported async validators #171 - [x] custom validation message for fields #169 - [x] supported string-based validators #167 - [x] vue-multiselect fixed #30 but in beta15, but they expanded [CSS from JS bundle](http://monterail.github.io/vue-multiselect/#sub-getting-started).
non_process
update docs there are some recent changes what we need to update in documentation too change property in schema of file input supported async validators custom validation message for fields supported string based validators vue multiselect fixed but in but they expanded
0
279,712
24,249,318,430
IssuesEvent
2022-09-27 13:08:16
hazelcast/hazelcast-python-client
https://api.github.com/repos/hazelcast/hazelcast-python-client
closed
test_translate_is_used
Type: Test-Failure Source: Internal
Failed on Windows against Python 2.7 https://github.com/hazelcast/hazelcast-python-client/runs/3889839024?check_suite_focus=true ``` ====================================================================== FAIL: test_translate_is_used (tests.integration.connection_manager_translate_test.ConnectionManagerTranslateTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "D:\a\hazelcast-python-client\hazelcast-python-client\tests\integration\connection_manager_translate_test.py", line 51, in test_translate_is_used connection_timeout=1.0, AssertionError: IllegalStateError not raised ```
1.0
test_translate_is_used - Failed on Windows against Python 2.7 https://github.com/hazelcast/hazelcast-python-client/runs/3889839024?check_suite_focus=true ``` ====================================================================== FAIL: test_translate_is_used (tests.integration.connection_manager_translate_test.ConnectionManagerTranslateTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "D:\a\hazelcast-python-client\hazelcast-python-client\tests\integration\connection_manager_translate_test.py", line 51, in test_translate_is_used connection_timeout=1.0, AssertionError: IllegalStateError not raised ```
non_process
test translate is used failed on windows against python fail test translate is used tests integration connection manager translate test connectionmanagertranslatetest traceback most recent call last file d a hazelcast python client hazelcast python client tests integration connection manager translate test py line in test translate is used connection timeout assertionerror illegalstateerror not raised
0
14,684
18,019,713,125
IssuesEvent
2021-09-16 17:47:08
lgblgblgb/xemu
https://api.github.com/repos/lgblgblgb/xemu
opened
MEGA65: V400 problem with non-bitplane modes?
target:MEGA65 compatibility-example VIC-IV
OpenROMs display only the top half of the screen, but not the bottom half, if 80x50 mode is used (it's OK 25 lines mode). Just trying to cursor down beyond the half of the screen, type something, and cursor up again: the cursor is again visible, the text is not, though in memory dump, the screen content clearly shows the typed text is there in the screen memory. According to FeralChild, 80x50 just works fine on the hardware. My theory: OpenROMs may use V400 to achieve the 50 lines resolution which is not handled correctly (or at all) by Xemu. This is maybe true for "legacy" VIC-II modes (hires, MCM?) as well, I've never tested that (I even don't know if it's supported to have double vertical resolution with V400 set then). I'm quite sure, this is handled in bitplane mode though, since it was tested with various BASIC programs from the ClosedROMs project. So my conclusion: this is maybe not handled in other modes too much. Note: `EFFICTIVE_V400` should be used instead of `REG_V400` macro because of some hacks which judges a certain situation as V200 even if it's V400.
True
MEGA65: V400 problem with non-bitplane modes? - OpenROMs display only the top half of the screen, but not the bottom half, if 80x50 mode is used (it's OK 25 lines mode). Just trying to cursor down beyond the half of the screen, type something, and cursor up again: the cursor is again visible, the text is not, though in memory dump, the screen content clearly shows the typed text is there in the screen memory. According to FeralChild, 80x50 just works fine on the hardware. My theory: OpenROMs may use V400 to achieve the 50 lines resolution which is not handled correctly (or at all) by Xemu. This is maybe true for "legacy" VIC-II modes (hires, MCM?) as well, I've never tested that (I even don't know if it's supported to have double vertical resolution with V400 set then). I'm quite sure, this is handled in bitplane mode though, since it was tested with various BASIC programs from the ClosedROMs project. So my conclusion: this is maybe not handled in other modes too much. Note: `EFFICTIVE_V400` should be used instead of `REG_V400` macro because of some hacks which judges a certain situation as V200 even if it's V400.
non_process
problem with non bitplane modes openroms display only the top half of the screen but not the bottom half if mode is used it s ok lines mode just trying to cursor down beyond the half of the screen type something and cursor up again the cursor is again visible the text is not though in memory dump the screen content clearly shows the typed text is there in the screen memory according to feralchild just works fine on the hardware my theory openroms may use to achieve the lines resolution which is not handled correctly or at all by xemu this is maybe true for legacy vic ii modes hires mcm as well i ve never tested that i even don t know if it s supported to have double vertical resolution with set then i m quite sure this is handled in bitplane mode though since it was tested with various basic programs from the closedroms project so my conclusion this is maybe not handled in other modes too much note effictive should be used instead of reg macro because of some hacks which judges a certain situation as even if it s
0
21,515
29,801,061,569
IssuesEvent
2023-06-16 08:08:43
openfoodfacts/openfoodfacts-server
https://api.github.com/repos/openfoodfacts/openfoodfacts-server
closed
The includes and their translations have not been deployed for Nova Groups pages
bug P1 static content nova processing Stale :star: top bug
### Describe the bug The Nova Group pages seem to not be loading right, regardless of the amount of time I've tried, they aren't loading especially the fresh one, lots of information is missing. ### To Reproduce https://fr.openfoodfacts.org/nova https://world.openfoodfacts.org/nova ### Expected behavior The pages should fully load and display the calculus methodology
1.0
The includes and their translations have not been deployed for Nova Groups pages - ### Describe the bug The Nova Group pages seem to not be loading right, regardless of the amount of time I've tried, they aren't loading especially the fresh one, lots of information is missing. ### To Reproduce https://fr.openfoodfacts.org/nova https://world.openfoodfacts.org/nova ### Expected behavior The pages should fully load and display the calculus methodology
process
the includes and their translations have not been deployed for nova groups pages describe the bug the nova group pages seem to not be loading right regardless of the amount of time i ve tried they aren t loading especially the fresh one lots of information is missing to reproduce expected behavior the pages should fully load and display the calculus methodology
1
21,825
30,316,774,062
IssuesEvent
2023-07-10 16:05:41
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
New Term - eventType
Term - add Class - Event normative Process - complete
## New term * Submitter: John Wieczorek * Efficacy Justification (why is this term necessary?): The hierarchical Event structure is not currently capable of distinguishing what different Events are for. * Demand Justification (name at least two organizations that independently need this term): The need for an eventType to distinguish distinct kinds of activities has been promoted in the Interactions Interest Group (https://github.com/tdwg/interaction/issues/22), the Camera Trap Data Package repository (https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931487715 and https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931504444), in discussions about the vocabulary for basisOfRecord (https://github.com/tdwg/dwc/issues/302), in the design of conceptual and data publishing models in Diversifying the GBIF Data Model Project (https://tinyurl.com/diversifying-gbif-data), and for [OBIS](https://github.com/iobis/obis-issues/issues/172) . * Stability Justification (what concerns are there that this might affect existing implementations?): The addition of this term will affect the definition of the GBIF Event Core, but this addition is already anticipated (see https://github.com/gbif/vocabulary/issues/107). * Implications for dwciri: namespace (does this change affect a dwciri term version)?: Yes, an equivalent term in the dwciri: namespace will be required. Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): eventType * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event * Definition of the term (normative): The nature of the Event. * Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use a controlled vocabulary. Regardless of the eventType, the interval of the Event should be captured in eventDate. * Examples (not normative): `Sample`, `Observation`, `Site Visit`, `Biotic Interaction`, `Bioblitz`, `Expedition`, `Survey`, `Project` (There is a prospective vocabulary for consideration at https://registry.gbif-uat.org/vocabulary/EventType/concepts, but this proposal is not to define an eventType vocabulary). * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): Not in ABCD
1.0
New Term - eventType - ## New term * Submitter: John Wieczorek * Efficacy Justification (why is this term necessary?): The hierarchical Event structure is not currently capable of distinguishing what different Events are for. * Demand Justification (name at least two organizations that independently need this term): The need for an eventType to distinguish distinct kinds of activities has been promoted in the Interactions Interest Group (https://github.com/tdwg/interaction/issues/22), the Camera Trap Data Package repository (https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931487715 and https://github.com/tdwg/camtrap-dp/issues/169#issuecomment-931504444), in discussions about the vocabulary for basisOfRecord (https://github.com/tdwg/dwc/issues/302), in the design of conceptual and data publishing models in Diversifying the GBIF Data Model Project (https://tinyurl.com/diversifying-gbif-data), and for [OBIS](https://github.com/iobis/obis-issues/issues/172) . * Stability Justification (what concerns are there that this might affect existing implementations?): The addition of this term will affect the definition of the GBIF Event Core, but this addition is already anticipated (see https://github.com/gbif/vocabulary/issues/107). * Implications for dwciri: namespace (does this change affect a dwciri term version)?: Yes, an equivalent term in the dwciri: namespace will be required. Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): eventType * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event * Definition of the term (normative): The nature of the Event. * Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use a controlled vocabulary. Regardless of the eventType, the interval of the Event should be captured in eventDate. * Examples (not normative): `Sample`, `Observation`, `Site Visit`, `Biotic Interaction`, `Bioblitz`, `Expedition`, `Survey`, `Project` (There is a prospective vocabulary for consideration at https://registry.gbif-uat.org/vocabulary/EventType/concepts, but this proposal is not to define an eventType vocabulary). * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): Not in ABCD
process
new term eventtype new term submitter john wieczorek efficacy justification why is this term necessary the hierarchical event structure is not currently capable of distinguishing what different events are for demand justification name at least two organizations that independently need this term the need for an eventtype to distinguish distinct kinds of activities has been promoted in the interactions interest group the camera trap data package repository and in discussions about the vocabulary for basisofrecord in the design of conceptual and data publishing models in diversifying the gbif data model project and for stability justification what concerns are there that this might affect existing implementations the addition of this term will affect the definition of the gbif event core but this addition is already anticipated see implications for dwciri namespace does this change affect a dwciri term version yes an equivalent term in the dwciri namespace will be required proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes eventtype organized in class e g occurrence event location taxon event definition of the term normative the nature of the event usage comments recommendations regarding content etc not normative recommended best practice is to use a controlled vocabulary regardless of the eventtype the interval of the event should be captured in eventdate examples not normative sample observation site visit biotic interaction bioblitz expedition survey project there is a prospective vocabulary for consideration at but this proposal is not to define an eventtype vocabulary refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
1
11,241
9,285,929,371
IssuesEvent
2019-03-21 08:58:14
astrolabsoftware/fink-broker
https://api.github.com/repos/astrolabsoftware/fink-broker
closed
Define common Apache Spark initialization step for services
apache kafka apache spark services
all services start with the same spark initialization steps (including setting the same Kafka parameters, etc). This is unnecessary redundancy, and prone to mistakes. The idea would be to factor this part, and include it into Fink core.
1.0
Define common Apache Spark initialization step for services - all services start with the same spark initialization steps (including setting the same Kafka parameters, etc). This is unnecessary redundancy, and prone to mistakes. The idea would be to factor this part, and include it into Fink core.
non_process
define common apache spark initialization step for services all services start with the same spark initialization steps including setting the same kafka parameters etc this is unnecessary redundancy and prone to mistakes the idea would be to factor this part and include it into fink core
0
166,064
14,018,951,902
IssuesEvent
2020-10-29 17:30:22
actions-pack/feedbacks
https://api.github.com/repos/actions-pack/feedbacks
closed
Compatibility with Advanced Custom Field Plugin?
documentation
![image](https://user-images.githubusercontent.com/17483092/77804436-80bce800-70a5-11ea-9f50-464d483fc396.png) You need to map your ACF field name in Registration Action's user meta fields. E.g one of the user meta field names is **phone** as shown in the image. So you just need to map your Elementor Form's Phone Field ID to **phone**
1.0
Compatibility with Advanced Custom Field Plugin? - ![image](https://user-images.githubusercontent.com/17483092/77804436-80bce800-70a5-11ea-9f50-464d483fc396.png) You need to map your ACF field name in Registration Action's user meta fields. E.g one of the user meta field names is **phone** as shown in the image. So you just need to map your Elementor Form's Phone Field ID to **phone**
non_process
compatibility with advanced custom field plugin you need to map your acf field name in registration action s user meta fields e g one of the user meta field names is phone as shown in the image so you just need to map your elementor form s phone field id to phone
0
12,186
9,594,825,905
IssuesEvent
2019-05-09 14:46:31
opencb/opencga
https://api.github.com/repos/opencb/opencga
opened
Upload web service fails to upload to root folder
bug web services
files/upload web service fails when the user attempts to upload files to the root folder of the path.
1.0
Upload web service fails to upload to root folder - files/upload web service fails when the user attempts to upload files to the root folder of the path.
non_process
upload web service fails to upload to root folder files upload web service fails when the user attempts to upload files to the root folder of the path
0
36,750
6,548,340,108
IssuesEvent
2017-09-04 20:58:09
ekeih/OmNomNom
https://api.github.com/repos/ekeih/OmNomNom
opened
Improve documentation
documentation
Currently the bot is missing mostly everything you would expect from a well documented project. If you are thinking about contributing code or just want to run the bot yourself, please let me know, so I can prioritize this issue ;-)
1.0
Improve documentation - Currently the bot is missing mostly everything you would expect from a well documented project. If you are thinking about contributing code or just want to run the bot yourself, please let me know, so I can prioritize this issue ;-)
non_process
improve documentation currently the bot is missing mostly everything you would expect from a well documented project if you are thinking about contributing code or just want to run the bot yourself please let me know so i can prioritize this issue
0
17,338
23,157,876,125
IssuesEvent
2022-07-29 14:38:52
Tencent/tdesign-miniprogram
https://api.github.com/repos/Tencent/tdesign-miniprogram
closed
[tabs] 希望支持选中的下标改色
good first issue in process
### 这个功能解决了什么问题 ![image](https://user-images.githubusercontent.com/57332732/181702966-348b5b0e-4993-4ad2-b68e-c672c7206ddc.png) 现在每次改色,都要手动去修改,极其不方便,希望能支持两种模式,纯改色,和slot的 ### 你建议的方案是什么 把设置单独做成组件属性
1.0
[tabs] 希望支持选中的下标改色 - ### 这个功能解决了什么问题 ![image](https://user-images.githubusercontent.com/57332732/181702966-348b5b0e-4993-4ad2-b68e-c672c7206ddc.png) 现在每次改色,都要手动去修改,极其不方便,希望能支持两种模式,纯改色,和slot的 ### 你建议的方案是什么 把设置单独做成组件属性
process
希望支持选中的下标改色 这个功能解决了什么问题 现在每次改色,都要手动去修改,极其不方便,希望能支持两种模式,纯改色,和slot的 你建议的方案是什么 把设置单独做成组件属性
1
5,335
8,154,440,234
IssuesEvent
2018-08-23 03:17:18
HumanCellAtlas/dcp-community
https://api.github.com/repos/HumanCellAtlas/dcp-community
closed
Objective or Objectives
charter-process
I think in the context of our charters Objective as found in https://github.com/HumanCellAtlas/dcp-community/blob/master/charters/charter-template.md tends to be used in the plural rather than singular Can this be updated to the plural Objectives?
1.0
Objective or Objectives - I think in the context of our charters Objective as found in https://github.com/HumanCellAtlas/dcp-community/blob/master/charters/charter-template.md tends to be used in the plural rather than singular Can this be updated to the plural Objectives?
process
objective or objectives i think in the context of our charters objective as found in tends to be used in the plural rather than singular can this be updated to the plural objectives
1
18,525
24,552,095,119
IssuesEvent
2022-10-12 13:21:37
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [Offline indicator] Offline feature is not working in the following scenario
Bug P1 iOS Process: Fixed Process: Tested dev
**Description** **Pre_condition:** The testing device should connect to WiFi **Steps:** 1. Install the mobile app on the testing device 2. Turn off the main data (Don't turn off the WI-FI of the testing device) 3. Observe **AR:** Offline indicator feature is not working **ER:** Offline indicator feature should work as per expected in the above scenario Note: Issue is not observed in android
2.0
[iOS] [Offline indicator] Offline feature is not working in the following scenario - **Description** **Pre_condition:** The testing device should connect to WiFi **Steps:** 1. Install the mobile app on the testing device 2. Turn off the main data (Don't turn off the WI-FI of the testing device) 3. Observe **AR:** Offline indicator feature is not working **ER:** Offline indicator feature should work as per expected in the above scenario Note: Issue is not observed in android
process
offline feature is not working in the following scenario description pre condition the testing device should connect to wifi steps install the mobile app on the testing device turn off the main data don t turn off the wi fi of the testing device observe ar offline indicator feature is not working er offline indicator feature should work as per expected in the above scenario note issue is not observed in android
1
1,703
4,349,928,413
IssuesEvent
2016-07-30 22:27:31
pwittchen/ReactiveSensors
https://api.github.com/repos/pwittchen/ReactiveSensors
closed
Release 0.1.2
release process
**Initial release notes**: - bumped RxJava dependency to v. 1.1.8 - bumped RxAndroid dependency to v. 1.2.1 - bumped Google Truth test dependency to v. 0.28 - bumped Compile SDK version to v. 23 - bumped Kotlin to v. 1.0.0 - updated sample apps **Things to do**: - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
1.0
Release 0.1.2 - **Initial release notes**: - bumped RxJava dependency to v. 1.1.8 - bumped RxAndroid dependency to v. 1.2.1 - bumped Google Truth test dependency to v. 0.28 - bumped Compile SDK version to v. 23 - bumped Kotlin to v. 1.0.0 - updated sample apps **Things to do**: - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
process
release initial release notes bumped rxjava dependency to v bumped rxandroid dependency to v bumped google truth test dependency to v bumped compile sdk version to v bumped kotlin to v updated sample apps things to do bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
1
427,697
12,397,948,738
IssuesEvent
2020-05-21 00:15:28
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
[osgi-cdi]OSGi service automatic publishing with @Publish-liking annotation
Component: OSGi-JavaEE ERR: Assignee Priority: Critical Stale Type: New Feature
Liking Weld-OSGi: allows developers to automatically publish service implementation.There is nothing to do, just put the annotation. OSGi framework is completely hidden. Then the service is accessible through CDI-OSGi service injection and OSGi classic mechanisms. In addition, on OSGi RFP-0146 Draft, CDI002 - The specification MUST make it possible to publish CDI beans in the OSGi Service Registry. So, this is a critical Requirement on CDI/OSGi Integration just as @OSGiService. #### Affected Versions [future release]
1.0
[osgi-cdi]OSGi service automatic publishing with @Publish-liking annotation - Liking Weld-OSGi: allows developers to automatically publish service implementation.There is nothing to do, just put the annotation. OSGi framework is completely hidden. Then the service is accessible through CDI-OSGi service injection and OSGi classic mechanisms. In addition, on OSGi RFP-0146 Draft, CDI002 - The specification MUST make it possible to publish CDI beans in the OSGi Service Registry. So, this is a critical Requirement on CDI/OSGi Integration just as @OSGiService. #### Affected Versions [future release]
non_process
osgi service automatic publishing with publish liking annotation liking weld osgi allows developers to automatically publish service implementation there is nothing to do just put the annotation osgi framework is completely hidden then the service is accessible through cdi osgi service injection and osgi classic mechanisms in addition on osgi rfp draft the specification must make it possible to publish cdi beans in the osgi service registry so this is a critical requirement on cdi osgi integration just as osgiservice affected versions
0
19,070
25,098,729,614
IssuesEvent
2022-11-08 12:07:55
hoprnet/hoprnet
https://api.github.com/repos/hoprnet/hoprnet
closed
Add staging branches for all supported releases
devops epic processes
We want to be able to merge PRs into release staging branches and only when we consider the sum of changes release-worthy, that branch is merged into the release branch. This process should be started with #4275 # Example Release branch: `release/bogota` Staging branch: `release-staging/bogota` 1. PRs are merged into `release-staging/bogota` 2. CI deployment workflow runs through and create Docker images and deploys nodes. 2.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2-next.3` 2.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota-staging` 3. Merge PR into branch `release/bogota` is created and merged 4. CI deployment workflow runs through and create Docker images and deploys nodes. 4.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2` 4.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota` # Requirements - [x] CI workflow is adapted to support execution on staging branches - [x] protocol config is adapted to support per release staging versions in the same environment - [x] version generation script, see `scripts/publish-version.sh` can generated staging release versions
1.0
Add staging branches for all supported releases - We want to be able to merge PRs into release staging branches and only when we consider the sum of changes release-worthy, that branch is merged into the release branch. This process should be started with #4275 # Example Release branch: `release/bogota` Staging branch: `release-staging/bogota` 1. PRs are merged into `release-staging/bogota` 2. CI deployment workflow runs through and create Docker images and deploys nodes. 2.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2-next.3` 2.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota-staging` 3. Merge PR into branch `release/bogota` is created and merged 4. CI deployment workflow runs through and create Docker images and deploys nodes. 4.1. Docker image name: `gcr.io/hoprassociation/hoprd:v1.91.2` 4.2. moving Docker tag: ``gcr.io/hoprassociation/hoprd:bogota` # Requirements - [x] CI workflow is adapted to support execution on staging branches - [x] protocol config is adapted to support per release staging versions in the same environment - [x] version generation script, see `scripts/publish-version.sh` can generated staging release versions
process
add staging branches for all supported releases we want to be able to merge prs into release staging branches and only when we consider the sum of changes release worthy that branch is merged into the release branch this process should be started with example release branch release bogota staging branch release staging bogota prs are merged into release staging bogota ci deployment workflow runs through and create docker images and deploys nodes docker image name gcr io hoprassociation hoprd next moving docker tag gcr io hoprassociation hoprd bogota staging merge pr into branch release bogota is created and merged ci deployment workflow runs through and create docker images and deploys nodes docker image name gcr io hoprassociation hoprd moving docker tag gcr io hoprassociation hoprd bogota requirements ci workflow is adapted to support execution on staging branches protocol config is adapted to support per release staging versions in the same environment version generation script see scripts publish version sh can generated staging release versions
1
2,598
5,356,200,627
IssuesEvent
2017-02-20 15:04:35
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Bootstrap Table Warnings
enhancement inprocess
Hi Allen, Thanks again for taking your time everyday to read through these issues. It is much appreciated. I'm getting the following warnings when I render your tables: ![image](https://cloud.githubusercontent.com/assets/5956893/22916224/29ec5f8c-f24b-11e6-8720-2ee08081da2f.png) This doesn't break the tables as of now, but is obviously a concern for the future. Currently using version 2.11.0.
1.0
Bootstrap Table Warnings - Hi Allen, Thanks again for taking your time everyday to read through these issues. It is much appreciated. I'm getting the following warnings when I render your tables: ![image](https://cloud.githubusercontent.com/assets/5956893/22916224/29ec5f8c-f24b-11e6-8720-2ee08081da2f.png) This doesn't break the tables as of now, but is obviously a concern for the future. Currently using version 2.11.0.
process
bootstrap table warnings hi allen thanks again for taking your time everyday to read through these issues it is much appreciated i m getting the following warnings when i render your tables this doesn t break the tables as of now but is obviously a concern for the future currently using version
1
57,767
24,223,083,146
IssuesEvent
2022-09-26 12:30:34
adi-H/top-picks
https://api.github.com/repos/adi-H/top-picks
closed
add change review data to product page
enhancement ui list-microservice ratings-microservice to be continued backend
continution of TBD **updated 27/8** ratings - [x] edit / update rating route + tests lists - [ ] find product in user lists ui - [ ] add ur own review - [ ] change ur review
2.0
add change review data to product page - continution of TBD **updated 27/8** ratings - [x] edit / update rating route + tests lists - [ ] find product in user lists ui - [ ] add ur own review - [ ] change ur review
non_process
add change review data to product page continution of tbd updated ratings edit update rating route tests lists find product in user lists ui add ur own review change ur review
0
2,670
5,468,638,854
IssuesEvent
2017-03-10 07:04:26
openslide/openslide
https://api.github.com/repos/openslide/openslide
closed
Audit security options pages
development-process enhancement
There are lots of new options for teams and collaboration now. We should go through and decide if any changes are needed. https://github.com/organizations/openslide/settings/member_privileges https://github.com/organizations/openslide/settings/oauth_application_policy https://github.com/openslide/openslide/settings/collaboration
1.0
Audit security options pages - There are lots of new options for teams and collaboration now. We should go through and decide if any changes are needed. https://github.com/organizations/openslide/settings/member_privileges https://github.com/organizations/openslide/settings/oauth_application_policy https://github.com/openslide/openslide/settings/collaboration
process
audit security options pages there are lots of new options for teams and collaboration now we should go through and decide if any changes are needed
1
337,745
30,259,979,870
IssuesEvent
2023-07-07 07:26:32
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: failover/non-system/pause failed
C-test-failure O-robot O-roachtest branch-master release-blocker T-kv
roachtest.failover/non-system/pause [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=artifacts#/failover/non-system/pause) on master @ [dbe8511fae8fca21562fdde5c240b1f7d06ef582](https://github.com/cockroachdb/cockroach/commits/dbe8511fae8fca21562fdde5c240b1f7d06ef582): ``` (test_runner.go:1075).runTest: test timed out (30m0s) (cluster.go:2282).Run: output in run_141100.736540940_n7_cockroach-workload-i: ./cockroach workload init kv --splits 1000 {pgurl:1} returned: COMMAND_PROBLEM: exit status 137 test artifacts and logs in: /artifacts/failover/non-system/pause/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=2</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/non-system/pause.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29488
2.0
roachtest: failover/non-system/pause failed - roachtest.failover/non-system/pause [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10797165?buildTab=artifacts#/failover/non-system/pause) on master @ [dbe8511fae8fca21562fdde5c240b1f7d06ef582](https://github.com/cockroachdb/cockroach/commits/dbe8511fae8fca21562fdde5c240b1f7d06ef582): ``` (test_runner.go:1075).runTest: test timed out (30m0s) (cluster.go:2282).Run: output in run_141100.736540940_n7_cockroach-workload-i: ./cockroach workload init kv --splits 1000 {pgurl:1} returned: COMMAND_PROBLEM: exit status 137 test artifacts and logs in: /artifacts/failover/non-system/pause/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=2</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/non-system/pause.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29488
non_process
roachtest failover non system pause failed roachtest failover non system pause with on master test runner go runtest test timed out cluster go run output in run cockroach workload i cockroach workload init kv splits pgurl returned command problem exit status test artifacts and logs in artifacts failover non system pause run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
0
473,818
13,648,059,297
IssuesEvent
2020-09-26 07:03:44
TerryCavanagh/diceydungeonsbeta
https://api.github.com/repos/TerryCavanagh/diceydungeonsbeta
closed
Precious Egg and similar should be immune to fury/have unique behavior
High Priority release candidate v0.1
When furied, precious egg duplicates, and this can make your skillcard inaccessible which is potentially run-breaking. Remix robot getting fury is really common so this can be a problem! Maybe it should just do nothing when used with fury, or skip a step? (Ex. precious egg 6 skips to 4, precious egg 2 just gives you the item inside)
1.0
Precious Egg and similar should be immune to fury/have unique behavior - When furied, precious egg duplicates, and this can make your skillcard inaccessible which is potentially run-breaking. Remix robot getting fury is really common so this can be a problem! Maybe it should just do nothing when used with fury, or skip a step? (Ex. precious egg 6 skips to 4, precious egg 2 just gives you the item inside)
non_process
precious egg and similar should be immune to fury have unique behavior when furied precious egg duplicates and this can make your skillcard inaccessible which is potentially run breaking remix robot getting fury is really common so this can be a problem maybe it should just do nothing when used with fury or skip a step ex precious egg skips to precious egg just gives you the item inside
0
2,007
4,827,340,176
IssuesEvent
2016-11-07 13:19:24
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Fix identifiers extraction in EUCTR
3. In Development bug Processors
Currently we just [prepend "EUCTR" to `eudract_number`](https://github.com/opentrials/processors/blob/master/processors/euctr/extractors.py#L27) but there are [records whose `eudract_number` looks like `EUCTR2014-001259-22-3rd`](https://query.opentrials.net/queries/140). They will end up [having no identifiers](https://explorer.opentrials.net/trials/42a1b0a4-8c52-11e6-a776-0242ac12000b). We should prepend smartly to avoid turning identifiers like the one above into invalid ones (`EUCTREUCTR2014-001259-22-3rd`). `euctr` processor is the one that has been observed but there may be others with this issue so they should all be checked.
1.0
Fix identifiers extraction in EUCTR - Currently we just [prepend "EUCTR" to `eudract_number`](https://github.com/opentrials/processors/blob/master/processors/euctr/extractors.py#L27) but there are [records whose `eudract_number` looks like `EUCTR2014-001259-22-3rd`](https://query.opentrials.net/queries/140). They will end up [having no identifiers](https://explorer.opentrials.net/trials/42a1b0a4-8c52-11e6-a776-0242ac12000b). We should prepend smartly to avoid turning identifiers like the one above into invalid ones (`EUCTREUCTR2014-001259-22-3rd`). `euctr` processor is the one that has been observed but there may be others with this issue so they should all be checked.
process
fix identifiers extraction in euctr currently we just but there are they will end up we should prepend smartly to avoid turning identifiers like the one above into invalid ones euctr processor is the one that has been observed but there may be others with this issue so they should all be checked
1
8,540
11,714,073,216
IssuesEvent
2020-03-09 11:34:44
kazuwjnlab/cvpaper
https://api.github.com/repos/kazuwjnlab/cvpaper
opened
[cvpaper] CVPR2019 #686 Libra R-CNN: Towards Balanced Learning for Object Detection
Object Detection imbalance training process
## \#686 [Libra R-CNN: Towards Balanced Learning for Object Detection](http://openaccess.thecvf.com/content_CVPR_2019/papers/Pang_Libra_R-CNN_Towards_Balanced_Learning_for_Object_Detection_CVPR_2019_paper.pdf) Jiangmiao Pang, Kai Chen, Jianping Shi, Huajun Feng, Wanli Ouyang, Dahua Lin ### どんな論文か? この研究は、物体検出の標準的な訓練方法を再検討します。 この論文では、検出性能についての言及は、トレーニングプロセス中の3つのレベルからなる不均衡によってしばしば制限されています。 1)サンプルレベル、2)特徴レベル、および3)客観的レベル。 この観察の結果として、本論文は天秤座R-CNN、物体検出のためのバランスのとれた学習に向けた簡単だが効果的なフレームワークを提案する。 ### 新規性 3つの要素を組み合わせることによってトレーニングプロセスのバランスを取り直すフレームワーク:a)IoUバランスサンプリング、追加費用なしのシンプルだが効果的なハードマイニング方法、b)同じ深く統合されたバランスのとれた意味機能を使用したマルチレベル機能の強化 、c)関連するサンプルおよびタスクをバランスを取り戻すための重大な回帰勾配を促進するための、L1損失の均衡化 ### 結果 この手法をMS COCOデータセットで評価し、最先端のオブジェクト検出手法と比較しました。 全体的にバランスのとれた設計により、提案されたLibra R-CNNは38.7 APを達成し、これはFPN Faster R-CNNよりも2.5ポイント高くなっています。 これは、はるかに強力な機能抽出機能であるResNetXt-101-64x4dで後者よりも優れています。 ### その他(なぜ通ったか?等) 著者らは彼らの研究を単一の検出器、Libra RetinaNet(Libra RN)に拡張した。 T.天秤座RPNからの大幅な改善。 結果を添付の表3に示す。 ![thumb](https://drive.google.com/uc?export=view&id=1NOFOgALLn-PR24B7wLrKIiF2UVLss5h4) object detection, training process, imbalance,
1.0
[cvpaper] CVPR2019 #686 Libra R-CNN: Towards Balanced Learning for Object Detection - ## \#686 [Libra R-CNN: Towards Balanced Learning for Object Detection](http://openaccess.thecvf.com/content_CVPR_2019/papers/Pang_Libra_R-CNN_Towards_Balanced_Learning_for_Object_Detection_CVPR_2019_paper.pdf) Jiangmiao Pang, Kai Chen, Jianping Shi, Huajun Feng, Wanli Ouyang, Dahua Lin ### どんな論文か? この研究は、物体検出の標準的な訓練方法を再検討します。 この論文では、検出性能についての言及は、トレーニングプロセス中の3つのレベルからなる不均衡によってしばしば制限されています。 1)サンプルレベル、2)特徴レベル、および3)客観的レベル。 この観察の結果として、本論文は天秤座R-CNN、物体検出のためのバランスのとれた学習に向けた簡単だが効果的なフレームワークを提案する。 ### 新規性 3つの要素を組み合わせることによってトレーニングプロセスのバランスを取り直すフレームワーク:a)IoUバランスサンプリング、追加費用なしのシンプルだが効果的なハードマイニング方法、b)同じ深く統合されたバランスのとれた意味機能を使用したマルチレベル機能の強化 、c)関連するサンプルおよびタスクをバランスを取り戻すための重大な回帰勾配を促進するための、L1損失の均衡化 ### 結果 この手法をMS COCOデータセットで評価し、最先端のオブジェクト検出手法と比較しました。 全体的にバランスのとれた設計により、提案されたLibra R-CNNは38.7 APを達成し、これはFPN Faster R-CNNよりも2.5ポイント高くなっています。 これは、はるかに強力な機能抽出機能であるResNetXt-101-64x4dで後者よりも優れています。 ### その他(なぜ通ったか?等) 著者らは彼らの研究を単一の検出器、Libra RetinaNet(Libra RN)に拡張した。 T.天秤座RPNからの大幅な改善。 結果を添付の表3に示す。 ![thumb](https://drive.google.com/uc?export=view&id=1NOFOgALLn-PR24B7wLrKIiF2UVLss5h4) object detection, training process, imbalance,
process
libra r cnn towards balanced learning for object detection jiangmiao pang kai chen jianping shi huajun feng wanli ouyang dahua lin どんな論文か? この研究は、物体検出の標準的な訓練方法を再検討します。 この論文では、検出性能についての言及は、 。 )サンプルレベル、 )特徴レベル、 )客観的レベル。 この観察の結果として、本論文は天秤座r cnn、物体検出のためのバランスのとれた学習に向けた簡単だが効果的なフレームワークを提案する。 新規性 :a)iouバランスサンプリング、追加費用なしのシンプルだが効果的なハードマイニング方法、b)同じ深く統合されたバランスのとれた意味機能を使用したマルチレベル機能の強化 、c)関連するサンプルおよびタスクをバランスを取り戻すための重大な回帰勾配を促進するための、 結果 この手法をms cocoデータセットで評価し、最先端のオブジェクト検出手法と比較しました。 全体的にバランスのとれた設計により、提案されたlibra r apを達成し、これはfpn faster r 。 これは、はるかに強力な機能抽出機能であるresnetxt 。 その他(なぜ通ったか?等) 著者らは彼らの研究を単一の検出器、libra retinanet(libra rn)に拡張した。 t 天秤座rpnからの大幅な改善。 。 object detection training process imbalance
1
21,780
30,294,577,725
IssuesEvent
2023-07-09 17:44:29
The-Data-Alchemists-Manipal/MindWave
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
closed
Eye Detection Using opencv
image-processing
## 💥 Proposal In this project, I am going to use open cv to detect eyes in the image/video.
1.0
Eye Detection Using opencv - ## 💥 Proposal In this project, I am going to use open cv to detect eyes in the image/video.
process
eye detection using opencv 💥 proposal in this project i am going to use open cv to detect eyes in the image video
1
57,139
8,142,320,778
IssuesEvent
2018-08-21 07:10:16
github/orchestrator
https://api.github.com/repos/github/orchestrator
closed
Differentiating GracefulTakeover from MasterFailure in Hook
documentation
I noticed that the Recovery Hooks do not seem to differentiate between planned failovers (Graceful-Master-Takeover) and actual failures. I had been having an issue where I has triggering a STONITH process in my PreFailoverProcesses and noticed that this was making my GracefulTakeovers fail. What was happening was I did not expect the Graceful Takeover to trigger the PreFailoverProcesses as there were seperate hooks for these so I assumed wrong. Thus, my STONITH process was triggering mid-takeover and shutting down the master. This is fine, though I think the documentation could better illuminate what each hook is for, except that I'm not actually sure how to differentiate the difference between planned takeovers and true failures. In a graceful takeover, I don't really want to STONITH my master as I trust the takeover process to handle the rotation (or maybe this is a bad assumption on my part?). I thought maybe ORC_FAILURE_TYPE ({failureType}) would work, but it indicates "DeadMaster". I checked the source to find the list of possible values and noticed there is a "NoProblem" state that doesn't appear to be used. Why does Orchestrator label a Takeover as "DeadMaster". I would expect something like "PlannedSwitch" or "Takeover" as my Master isn't really dead. Anyways, I tried looking in the source that generates the magic replacement variables and the environment variables and noticed one that is not documented in the md files: ORC_COMMAND. I added this to my hooks and viola: "ORC_COMMAND: graceful-master-takeover". Excellent! I could pass this value to my STONITH process and have it either run a system check or do nothing. However, since it's undocumented, i'm not sure if it's wise to use it. Is this just a missing documentation item, or should I not be using this variable in this manner? Thanks
1.0
Differentiating GracefulTakeover from MasterFailure in Hook - I noticed that the Recovery Hooks do not seem to differentiate between planned failovers (Graceful-Master-Takeover) and actual failures. I had been having an issue where I has triggering a STONITH process in my PreFailoverProcesses and noticed that this was making my GracefulTakeovers fail. What was happening was I did not expect the Graceful Takeover to trigger the PreFailoverProcesses as there were seperate hooks for these so I assumed wrong. Thus, my STONITH process was triggering mid-takeover and shutting down the master. This is fine, though I think the documentation could better illuminate what each hook is for, except that I'm not actually sure how to differentiate the difference between planned takeovers and true failures. In a graceful takeover, I don't really want to STONITH my master as I trust the takeover process to handle the rotation (or maybe this is a bad assumption on my part?). I thought maybe ORC_FAILURE_TYPE ({failureType}) would work, but it indicates "DeadMaster". I checked the source to find the list of possible values and noticed there is a "NoProblem" state that doesn't appear to be used. Why does Orchestrator label a Takeover as "DeadMaster". I would expect something like "PlannedSwitch" or "Takeover" as my Master isn't really dead. Anyways, I tried looking in the source that generates the magic replacement variables and the environment variables and noticed one that is not documented in the md files: ORC_COMMAND. I added this to my hooks and viola: "ORC_COMMAND: graceful-master-takeover". Excellent! I could pass this value to my STONITH process and have it either run a system check or do nothing. However, since it's undocumented, i'm not sure if it's wise to use it. Is this just a missing documentation item, or should I not be using this variable in this manner? Thanks
non_process
differentiating gracefultakeover from masterfailure in hook i noticed that the recovery hooks do not seem to differentiate between planned failovers graceful master takeover and actual failures i had been having an issue where i has triggering a stonith process in my prefailoverprocesses and noticed that this was making my gracefultakeovers fail what was happening was i did not expect the graceful takeover to trigger the prefailoverprocesses as there were seperate hooks for these so i assumed wrong thus my stonith process was triggering mid takeover and shutting down the master this is fine though i think the documentation could better illuminate what each hook is for except that i m not actually sure how to differentiate the difference between planned takeovers and true failures in a graceful takeover i don t really want to stonith my master as i trust the takeover process to handle the rotation or maybe this is a bad assumption on my part i thought maybe orc failure type failuretype would work but it indicates deadmaster i checked the source to find the list of possible values and noticed there is a noproblem state that doesn t appear to be used why does orchestrator label a takeover as deadmaster i would expect something like plannedswitch or takeover as my master isn t really dead anyways i tried looking in the source that generates the magic replacement variables and the environment variables and noticed one that is not documented in the md files orc command i added this to my hooks and viola orc command graceful master takeover excellent i could pass this value to my stonith process and have it either run a system check or do nothing however since it s undocumented i m not sure if it s wise to use it is this just a missing documentation item or should i not be using this variable in this manner thanks
0
7,229
10,368,257,304
IssuesEvent
2019-09-07 15:29:55
banctilrobitaille/kerosene
https://api.github.com/repos/banctilrobitaille/kerosene
closed
[FEATURE] Implement a standard api for custom variable ploting
EventPreprocessor ploting
**Is your feature request related to a problem? Please describe.** No. **Describe the solution you'd like** Instead of having different EventPreprocessor for every visdom plot type we should have a single event preprocessor for custom variables that accepts the type plot, frequency and opts as args. **Describe alternatives you've considered** none **Will this change the current api? How?** Yes, instead of having multiple Preprocessor for custom variable there will be one. **Additional context** No.
1.0
[FEATURE] Implement a standard api for custom variable ploting - **Is your feature request related to a problem? Please describe.** No. **Describe the solution you'd like** Instead of having different EventPreprocessor for every visdom plot type we should have a single event preprocessor for custom variables that accepts the type plot, frequency and opts as args. **Describe alternatives you've considered** none **Will this change the current api? How?** Yes, instead of having multiple Preprocessor for custom variable there will be one. **Additional context** No.
process
implement a standard api for custom variable ploting is your feature request related to a problem please describe no describe the solution you d like instead of having different eventpreprocessor for every visdom plot type we should have a single event preprocessor for custom variables that accepts the type plot frequency and opts as args describe alternatives you ve considered none will this change the current api how yes instead of having multiple preprocessor for custom variable there will be one additional context no
1
69,926
30,505,008,875
IssuesEvent
2023-07-18 16:13:59
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
Error: could not find DNS zone
bug service/dns
### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.3.9 ### AzureRM Provider Version 3.46.0 ### Affected Resource(s)/Data Source(s) azurerm_dns_zone ### Terraform Configuration Files ```hcl data "azurerm_dns_zone" "this" { name = var.domain.name } ``` ### Debug Output/Panic Output ```shell ╷ │ Error: could not find DNS zone with name: "foo.com" │ │ with module.cdn[0].data.azurerm_dns_zone.this[0], │ on modules/cdn/cdn.tf line 44, in data "azurerm_dns_zone" "this": │ 44: data "azurerm_dns_zone" "this" { │ ╵ ``` ### Expected Behaviour Work ### Actual Behaviour Doesn't work. ### Steps to Reproduce _No response_ ### Important Factoids It works with hashicorp/azurerm = 3.45.0 and earlier. ### References _No response_
1.0
Error: could not find DNS zone - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.3.9 ### AzureRM Provider Version 3.46.0 ### Affected Resource(s)/Data Source(s) azurerm_dns_zone ### Terraform Configuration Files ```hcl data "azurerm_dns_zone" "this" { name = var.domain.name } ``` ### Debug Output/Panic Output ```shell ╷ │ Error: could not find DNS zone with name: "foo.com" │ │ with module.cdn[0].data.azurerm_dns_zone.this[0], │ on modules/cdn/cdn.tf line 44, in data "azurerm_dns_zone" "this": │ 44: data "azurerm_dns_zone" "this" { │ ╵ ``` ### Expected Behaviour Work ### Actual Behaviour Doesn't work. ### Steps to Reproduce _No response_ ### Important Factoids It works with hashicorp/azurerm = 3.45.0 and earlier. ### References _No response_
non_process
error could not find dns zone is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s azurerm dns zone terraform configuration files hcl data azurerm dns zone this name var domain name debug output panic output shell ╷ │ error could not find dns zone with name foo com │ │ with module cdn data azurerm dns zone this │ on modules cdn cdn tf line in data azurerm dns zone this │ data azurerm dns zone this │ ╵ expected behaviour work actual behaviour doesn t work steps to reproduce no response important factoids it works with hashicorp azurerm and earlier references no response
0
9,944
11,948,421,548
IssuesEvent
2020-04-03 11:49:57
icatproject/python-icat
https://api.github.com/repos/icatproject/python-icat
opened
Consider to switch to suds-community
compatibility
The original version of `suds` is dead since very long time. Jurko Gospodnetić has been taken over and created the `suds-jurko` fork. But development seem to have stalled again since 2015. Many new forks have been created since then, most of them rather short-lived. Now, there is one that lasted a little bit longer and at least has recent activity: [suds-community](https://pypi.org/project/suds-community/). Should test the latter and evaluate whether it works with python-icat. If positive, update documentation to point to that one as system requirement in the README.
True
Consider to switch to suds-community - The original version of `suds` is dead since very long time. Jurko Gospodnetić has been taken over and created the `suds-jurko` fork. But development seem to have stalled again since 2015. Many new forks have been created since then, most of them rather short-lived. Now, there is one that lasted a little bit longer and at least has recent activity: [suds-community](https://pypi.org/project/suds-community/). Should test the latter and evaluate whether it works with python-icat. If positive, update documentation to point to that one as system requirement in the README.
non_process
consider to switch to suds community the original version of suds is dead since very long time jurko gospodnetić has been taken over and created the suds jurko fork but development seem to have stalled again since many new forks have been created since then most of them rather short lived now there is one that lasted a little bit longer and at least has recent activity should test the latter and evaluate whether it works with python icat if positive update documentation to point to that one as system requirement in the readme
0
729,081
25,108,544,443
IssuesEvent
2022-11-08 18:30:10
cse-sim/cse
https://api.github.com/repos/cse-sim/cse
opened
Review and consider localizing environment-specific code
1 - low priority
The original idea of envpak was to isolate OS-specific and/or compiler-specific code -- provide a uniform internal API for the associated features. There are OS and compiler dependencies cropping up in other files. Notably rmkerr, but others also. Those dependencies could be reviewed and some capabilities moved to envpak as sensible / practical.
1.0
Review and consider localizing environment-specific code - The original idea of envpak was to isolate OS-specific and/or compiler-specific code -- provide a uniform internal API for the associated features. There are OS and compiler dependencies cropping up in other files. Notably rmkerr, but others also. Those dependencies could be reviewed and some capabilities moved to envpak as sensible / practical.
non_process
review and consider localizing environment specific code the original idea of envpak was to isolate os specific and or compiler specific code provide a uniform internal api for the associated features there are os and compiler dependencies cropping up in other files notably rmkerr but others also those dependencies could be reviewed and some capabilities moved to envpak as sensible practical
0
611,590
18,959,053,651
IssuesEvent
2021-11-19 00:51:38
GoogleChrome/lighthouse
https://api.github.com/repos/GoogleChrome/lighthouse
closed
Lighthouse test feature in Chrome just stopped working
bug needs-priority
### FAQ - [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/master/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/master/docs/throttling.md). - [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead). ### URL https://m.warsztat24.pl ### What happened? I cannot perform lighthouse test on https://m.warsztat24.pl for a few days now... there wasn't any problems before... but now... test runs and runs and runs (messages on popup do change all over the time) and just cannot finish and go to the summary.... what can be the reason? please help! ### What did you expect? I would like just to get lighthouse test results for https://m.warsztat24.pl. ### What have you tried? don't know what could I do, as the tests seem continue over and over... ### How were you running Lighthouse? Chrome DevTools ### Lighthouse Version don't know because I cannot see any finished report ### Chrome Version 94.0.4606.81 ### Node Version 12.16.1 ### Relevant log output _No response_
1.0
Lighthouse test feature in Chrome just stopped working - ### FAQ - [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/master/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/master/docs/throttling.md). - [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead). ### URL https://m.warsztat24.pl ### What happened? I cannot perform lighthouse test on https://m.warsztat24.pl for a few days now... there wasn't any problems before... but now... test runs and runs and runs (messages on popup do change all over the time) and just cannot finish and go to the summary.... what can be the reason? please help! ### What did you expect? I would like just to get lighthouse test results for https://m.warsztat24.pl. ### What have you tried? don't know what could I do, as the tests seem continue over and over... ### How were you running Lighthouse? Chrome DevTools ### Lighthouse Version don't know because I cannot see any finished report ### Chrome Version 94.0.4606.81 ### Node Version 12.16.1 ### Relevant log output _No response_
non_process
lighthouse test feature in chrome just stopped working faq yes my issue is not about or yes my issue is not about a specific accessibility audit file with instead url what happened i cannot perform lighthouse test on for a few days now there wasn t any problems before but now test runs and runs and runs messages on popup do change all over the time and just cannot finish and go to the summary what can be the reason please help what did you expect i would like just to get lighthouse test results for what have you tried don t know what could i do as the tests seem continue over and over how were you running lighthouse chrome devtools lighthouse version don t know because i cannot see any finished report chrome version node version relevant log output no response
0
10,147
13,044,162,542
IssuesEvent
2020-07-29 03:47:33
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `Sleep` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `Sleep` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @breeswish ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `Sleep` from TiDB - ## Description Port the scalar function `Sleep` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @breeswish ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function sleep from tidb description port the scalar function sleep from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
1
372,775
26,019,138,222
IssuesEvent
2022-12-21 11:02:37
boostercloud/booster
https://api.github.com/repos/boostercloud/booster
opened
Write documentation on how to extend existing provider implementations or create your own from scratch
documentation provider:multicloud
It would be extremely helpful to have documentation on extending existing provider implementations or creating new ones from scratch. This would allow developers to customize the infrastructure to meet their specific requirements more easily, and it would also make Booster easier to adopt in specific environments with special compliance requirements. Currently, there is very little information available on how to do this, as it would require developers to reverse engineering existing implementations, and it can be difficult for developers to figure it out on their own. This lack of documentation has likely been a barrier to adoption for many people.
1.0
Write documentation on how to extend existing provider implementations or create your own from scratch - It would be extremely helpful to have documentation on extending existing provider implementations or creating new ones from scratch. This would allow developers to customize the infrastructure to meet their specific requirements more easily, and it would also make Booster easier to adopt in specific environments with special compliance requirements. Currently, there is very little information available on how to do this, as it would require developers to reverse engineering existing implementations, and it can be difficult for developers to figure it out on their own. This lack of documentation has likely been a barrier to adoption for many people.
non_process
write documentation on how to extend existing provider implementations or create your own from scratch it would be extremely helpful to have documentation on extending existing provider implementations or creating new ones from scratch this would allow developers to customize the infrastructure to meet their specific requirements more easily and it would also make booster easier to adopt in specific environments with special compliance requirements currently there is very little information available on how to do this as it would require developers to reverse engineering existing implementations and it can be difficult for developers to figure it out on their own this lack of documentation has likely been a barrier to adoption for many people
0
7,271
10,425,352,520
IssuesEvent
2019-09-16 15:15:50
SpongePowered/Mixin
https://api.github.com/repos/SpongePowered/Mixin
closed
Add support for SuppressWarnings annotation in Mixin AP to allow silencing nuisance warnings
annotation processor enhancement
I want to have zero warnings when compiling my code, but currently mixin makes this impossible. This is basically the same issue as #290 but since `@Dynamic` not helping is intentional, it was closed. But this still leaves no solution to silence the warnings. Consider there is a class `X` (in my case `RenderChunk`) that some coremod (in my case, OptiFine) modifies by adding some fields (`renderChunkNeighboursValid` and `renderChunkNeighbours`), that I need to access within that my mixin. Adding a `@Shadow` for these fields works perfectly fine when running the code, but mixin AP will show the following error, even with `remap=false`: ``` MixinRenderChunk.java:51: warning: Cannot find target for @Shadow field in net.minecraft.client.renderer.chunk.RenderChunk @Dynamic @Shadow(remap = false) private RenderChunk[] renderChunkNeighbours; ^ ``` While it may not look like a big issue ("it's just warnings"), it makes it very hard to notice the meaningful warnings that I don't want to ignore in the compile output.
1.0
Add support for SuppressWarnings annotation in Mixin AP to allow silencing nuisance warnings - I want to have zero warnings when compiling my code, but currently mixin makes this impossible. This is basically the same issue as #290 but since `@Dynamic` not helping is intentional, it was closed. But this still leaves no solution to silence the warnings. Consider there is a class `X` (in my case `RenderChunk`) that some coremod (in my case, OptiFine) modifies by adding some fields (`renderChunkNeighboursValid` and `renderChunkNeighbours`), that I need to access within that my mixin. Adding a `@Shadow` for these fields works perfectly fine when running the code, but mixin AP will show the following error, even with `remap=false`: ``` MixinRenderChunk.java:51: warning: Cannot find target for @Shadow field in net.minecraft.client.renderer.chunk.RenderChunk @Dynamic @Shadow(remap = false) private RenderChunk[] renderChunkNeighbours; ^ ``` While it may not look like a big issue ("it's just warnings"), it makes it very hard to notice the meaningful warnings that I don't want to ignore in the compile output.
process
add support for suppresswarnings annotation in mixin ap to allow silencing nuisance warnings i want to have zero warnings when compiling my code but currently mixin makes this impossible this is basically the same issue as but since dynamic not helping is intentional it was closed but this still leaves no solution to silence the warnings consider there is a class x in my case renderchunk that some coremod in my case optifine modifies by adding some fields renderchunkneighboursvalid and renderchunkneighbours that i need to access within that my mixin adding a shadow for these fields works perfectly fine when running the code but mixin ap will show the following error even with remap false mixinrenderchunk java warning cannot find target for shadow field in net minecraft client renderer chunk renderchunk dynamic shadow remap false private renderchunk renderchunkneighbours while it may not look like a big issue it s just warnings it makes it very hard to notice the meaningful warnings that i don t want to ignore in the compile output
1
767,487
26,927,623,433
IssuesEvent
2023-02-07 14:48:17
daisy/ebraille
https://api.github.com/repos/daisy/ebraille
opened
Automatic, converted, and human prepared files will have different levels of quality and expectations
use case High Priority content spec metadata
I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format. *Detail* The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format. *Proposal* Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion. With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
1.0
Automatic, converted, and human prepared files will have different levels of quality and expectations - I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format. *Detail* The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format. *Proposal* Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion. With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
non_process
automatic converted and human prepared files will have different levels of quality and expectations i am a braille user and my expectations for the quality of file i am receiving will vary based on whether the file was prepared by a transcriber an automatic process or converted from a braille file in an older format detail the best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber next will likely be those converted from an automatic process finally there will be files prepared by converting a braille file in an older format proposal metadata should indicate whether the file was prepared by a human an automatic process or converted from an older braille file if a human it would be good if it indicated the person s level of braille competence professional vs amateur though this aspect needs more thought if an automatic process it would be good if it indicated the file type of the source file as each will have strengths and weaknesses the program used for the automatic process would also be good to indicate finally files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion with this information schools libraries and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes programs and starting files work best for them and their needs
0
340,408
10,272,042,724
IssuesEvent
2019-08-23 15:27:18
KSP-SpaceDock/SpaceDock
https://api.github.com/repos/KSP-SpaceDock/SpaceDock
opened
On the Game page the banner "Browse N more Mods" counts unpublished mods
Priority: Low Type: Backend Type: Bug
Description says it all
1.0
On the Game page the banner "Browse N more Mods" counts unpublished mods - Description says it all
non_process
on the game page the banner browse n more mods counts unpublished mods description says it all
0
6,944
10,112,666,444
IssuesEvent
2019-07-30 15:07:08
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
PowerShell Steps for ML Anomaly Detection
Pri2 assigned-to-author machine-learning/svc product-question team-data-science-process/subsvc triaged
Will the 'Invoke-MLAnomalyapi', as shown in the below video, be added to the Az.MachineLearning module in future? https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443 [![PS AI automation](https://img.youtube.com/vi/1EVHChiqZOw/0.jpg)](https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2d8a13e5-13f6-ffa2-ba7a-61a04c278293 * Version Independent ID: bbde9854-2484-e8aa-0eb8-8b641706a265 * Content: [Azure Machine Learning Anomaly Detection API - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/apps-anomaly-detection-api) * Content Source: [articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md](https://github.com/Microsoft/azure-docs/blob/master/articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md) * Service: **machine-learning** * Sub-service: **team-data-science-process** * GitHub Login: @marktab * Microsoft Alias: **tdsp**
1.0
PowerShell Steps for ML Anomaly Detection - Will the 'Invoke-MLAnomalyapi', as shown in the below video, be added to the Az.MachineLearning module in future? https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443 [![PS AI automation](https://img.youtube.com/vi/1EVHChiqZOw/0.jpg)](https://www.youtube.com/embed/1EVHChiqZOw?start=1399&end=1443) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2d8a13e5-13f6-ffa2-ba7a-61a04c278293 * Version Independent ID: bbde9854-2484-e8aa-0eb8-8b641706a265 * Content: [Azure Machine Learning Anomaly Detection API - Team Data Science Process](https://docs.microsoft.com/en-us/azure/machine-learning/team-data-science-process/apps-anomaly-detection-api) * Content Source: [articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md](https://github.com/Microsoft/azure-docs/blob/master/articles/machine-learning/team-data-science-process/apps-anomaly-detection-api.md) * Service: **machine-learning** * Sub-service: **team-data-science-process** * GitHub Login: @marktab * Microsoft Alias: **tdsp**
process
powershell steps for ml anomaly detection will the invoke mlanomalyapi as shown in the below video be added to the az machinelearning module in future document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service machine learning sub service team data science process github login marktab microsoft alias tdsp
1
11,824
14,648,348,865
IssuesEvent
2020-12-27 02:13:47
YDongY/Coding
https://api.github.com/repos/YDongY/Coding
opened
02 Process Control | Coding
/docs/Shell/ShellScript/02-ProcessControl/ Gitalk
https://coding.ydongy.cn/docs/Shell/ShellScript/02-ProcessControl/ 流程控制 # 条件判断 # if-then # if 条件/命令; then 指令 fi bash shell 的 if 语句会运行 if 后面的那个命令。如果该命令的退出状态码是 0(该命令成功运行),位于 then 部分的命令就会被执行。如果该命令的退出状态码是其他值,then 部分的命令就不会被执行。fi 语句用来表示 if-then 语句到此结束 #!/bin/bash if pwd ; then echo $(pwd) fi if-then-else # if 条件 ; then 指令1 else 指令2 fi if pwd ; then echo "success." else echo "fail." fi if-elif-then-else # if 条件 then 指令1 elif 条件2 then 指令2 else 指令3 fi #!
1.0
02 Process Control | Coding - https://coding.ydongy.cn/docs/Shell/ShellScript/02-ProcessControl/ 流程控制 # 条件判断 # if-then # if 条件/命令; then 指令 fi bash shell 的 if 语句会运行 if 后面的那个命令。如果该命令的退出状态码是 0(该命令成功运行),位于 then 部分的命令就会被执行。如果该命令的退出状态码是其他值,then 部分的命令就不会被执行。fi 语句用来表示 if-then 语句到此结束 #!/bin/bash if pwd ; then echo $(pwd) fi if-then-else # if 条件 ; then 指令1 else 指令2 fi if pwd ; then echo "success." else echo "fail." fi if-elif-then-else # if 条件 then 指令1 elif 条件2 then 指令2 else 指令3 fi #!
process
process control coding 流程控制 条件判断 if then if 条件 命令 then 指令 fi bash shell 的 if 语句会运行 if 后面的那个命令。如果该命令的退出状态码是 (该命令成功运行),位于 then 部分的命令就会被执行。如果该命令的退出状态码是其他值,then 部分的命令就不会被执行。fi 语句用来表示 if then 语句到此结束 bin bash if pwd then echo pwd fi if then else if 条件 then else fi if pwd then echo success else echo fail fi if elif then else if 条件 then elif then else fi
1
678,133
23,189,733,062
IssuesEvent
2022-08-01 11:34:29
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
opened
[News in administration tab] Admin can not change information in any news item without changing the date of the news to the current date or future date
bug Backend Priority: Medium
**Environment:** Windows 10 Professional, Chrome 96.0.4664.110 **Reproducible:** always **Build found:** https://speak-ukrainian.org.ua/dev/ **Preconditions** 1. Log in as an administrator on https://speak-ukrainian.org.ua/dev/ (login: [admin@gmail.com](mailto:admin@gmail.com); password: admin) 2. Go to the administration tab - "Новини" **Steps to reproduce** 1. Click on "Редагувати" button of any news item 3. Change something into current news item and press Save 4. Pay attention to the result of this action **Actual result** After pressing Save button an error "date Дата повинна бути сьогоднішньою або майбутньою." occurs **Expected result** Admin can change any news item without changing the date of the news ![помилка редагування новини](https://user-images.githubusercontent.com/42620939/182139288-e53d1786-4e54-4dbe-bf4a-e6798eeb93af.png)
1.0
[News in administration tab] Admin can not change information in any news item without changing the date of the news to the current date or future date - **Environment:** Windows 10 Professional, Chrome 96.0.4664.110 **Reproducible:** always **Build found:** https://speak-ukrainian.org.ua/dev/ **Preconditions** 1. Log in as an administrator on https://speak-ukrainian.org.ua/dev/ (login: [admin@gmail.com](mailto:admin@gmail.com); password: admin) 2. Go to the administration tab - "Новини" **Steps to reproduce** 1. Click on "Редагувати" button of any news item 3. Change something into current news item and press Save 4. Pay attention to the result of this action **Actual result** After pressing Save button an error "date Дата повинна бути сьогоднішньою або майбутньою." occurs **Expected result** Admin can change any news item without changing the date of the news ![помилка редагування новини](https://user-images.githubusercontent.com/42620939/182139288-e53d1786-4e54-4dbe-bf4a-e6798eeb93af.png)
non_process
admin can not change information in any news item without changing the date of the news to the current date or future date environment windows professional chrome reproducible always build found preconditions log in as an administrator on login mailto admin gmail com password admin go to the administration tab новини steps to reproduce click on редагувати button of any news item change something into current news item and press save pay attention to the result of this action actual result after pressing save button an error date дата повинна бути сьогоднішньою або майбутньою occurs expected result admin can change any news item without changing the date of the news
0
5,435
8,299,549,511
IssuesEvent
2018-09-21 03:35:31
flutterchina/dio
https://api.github.com/repos/flutterchina/dio
closed
Content size exceeds specified contentLength
processing
## Steps to Reproduce - 请求方式 ```java Map<String, dynamic> data dio.post(path, data: data); ``` - 请求失败 ```java requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: 测试测试测试, courseId: 110, score: 3, anonymous: 1, evaluateObject: 2, timestamp: 1537426578634} ``` 如果我把content换成英文,则请求成功。 - 请求成功 ```java requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: test, courseId: 110, score: 5, anonymous: 1, evaluateObject: 1, timestamp: 1537426565722} ``` ## Logs ``` DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"测试测试测试","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}] I/flutter (14570): DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"测试测试测试","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}] ```
1.0
Content size exceeds specified contentLength - ## Steps to Reproduce - 请求方式 ```java Map<String, dynamic> data dio.post(path, data: data); ``` - 请求失败 ```java requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: 测试测试测试, courseId: 110, score: 3, anonymous: 1, evaluateObject: 2, timestamp: 1537426578634} ``` 如果我把content换成英文,则请求成功。 - 请求成功 ```java requestPost host:http://api-shuangshi.test.tengyue360.com/ path:/backend/student/stuApp/evaluate data:{content: test, courseId: 110, score: 5, anonymous: 1, evaluateObject: 1, timestamp: 1537426565722} ``` ## Logs ``` DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"测试测试测试","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}] I/flutter (14570): DioError [DioErrorType.DEFAULT]: HttpException: Content size exceeds specified contentLength. 124 bytes written while expected 112. [{"content":"测试测试测试","courseId":"110","score":3,"anonymous":"1","evaluateObject":"2","timestamp":"1537426578634"}] ```
process
content size exceeds specified contentlength steps to reproduce 请求方式 java map data dio post path data data 请求失败 java requestpost host path backend student stuapp evaluate data content 测试测试测试 courseid score anonymous evaluateobject timestamp 如果我把content换成英文,则请求成功。 请求成功 java requestpost host path backend student stuapp evaluate data content test courseid score anonymous evaluateobject timestamp logs dioerror httpexception content size exceeds specified contentlength bytes written while expected i flutter dioerror httpexception content size exceeds specified contentlength bytes written while expected
1
9,992
13,039,523,833
IssuesEvent
2020-07-28 16:54:01
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
[tests] Tests in tables/automl/dataset_test.py expect to have access to gs://python-docs-samples-tests-automl-tables-test
api: automl priority: p2 type: process
## In which file did you encounter the issue? tables/automl/dataset_test.py ## Describe the issue All tests are failing with error: AccessDeniedException: 403 xxx@xxx.com does not have storage.objects.list access to the Google Cloud Storage bucket. It's not clear what account should I use for tests. All that said in AUTHORING_GUIDE.MD is that I need to setup my own project. But this GCS absolute bucket path is hard-coded inside tests file. How should I get access to this bucket?
1.0
[tests] Tests in tables/automl/dataset_test.py expect to have access to gs://python-docs-samples-tests-automl-tables-test - ## In which file did you encounter the issue? tables/automl/dataset_test.py ## Describe the issue All tests are failing with error: AccessDeniedException: 403 xxx@xxx.com does not have storage.objects.list access to the Google Cloud Storage bucket. It's not clear what account should I use for tests. All that said in AUTHORING_GUIDE.MD is that I need to setup my own project. But this GCS absolute bucket path is hard-coded inside tests file. How should I get access to this bucket?
process
tests in tables automl dataset test py expect to have access to gs python docs samples tests automl tables test in which file did you encounter the issue tables automl dataset test py describe the issue all tests are failing with error accessdeniedexception xxx xxx com does not have storage objects list access to the google cloud storage bucket it s not clear what account should i use for tests all that said in authoring guide md is that i need to setup my own project but this gcs absolute bucket path is hard coded inside tests file how should i get access to this bucket
1
205,136
23,299,717,965
IssuesEvent
2022-08-07 05:57:10
TheFloAnd/eventlisting-laravel
https://api.github.com/repos/TheFloAnd/eventlisting-laravel
closed
laravel-mix-6.0.43.tgz: 2 vulnerabilities (highest severity is: 7.5) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel-mix-6.0.43.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/terser/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-25858](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | detected in multiple dependencies | Transitive | N/A | &#10060; | | [CVE-2021-35065](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-5.1.2.tgz | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25858</summary> ### Vulnerable Libraries - <b>terser-4.8.0.tgz</b>, <b>terser-5.12.1.tgz</b></p> <p> ### <b>terser-4.8.0.tgz</b></p> <p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p> <p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-4.8.0.tgz">https://registry.npmjs.org/terser/-/terser-4.8.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/html-minifier-terser/node_modules/terser/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.43.tgz (Root Library) - html-loader-1.3.2.tgz - html-minifier-terser-5.1.1.tgz - :x: **terser-4.8.0.tgz** (Vulnerable Library) ### <b>terser-5.12.1.tgz</b></p> <p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p> <p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-5.12.1.tgz">https://registry.npmjs.org/terser/-/terser-5.12.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/terser/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.43.tgz (Root Library) - :x: **terser-5.12.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The package terser before 4.8.1, from 5.0.0 and before 5.14.2 are vulnerable to Regular Expression Denial of Service (ReDoS) due to insecure usage of regular expressions. <p>Publish Date: 2022-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858>CVE-2022-25858</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858</a></p> <p>Release Date: 2022-07-15</p> <p>Fix Resolution: terser - 4.8.1,5.14.2</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-35065</summary> ### Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.43.tgz (Root Library) - chokidar-3.5.3.tgz - :x: **glob-parent-5.1.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
laravel-mix-6.0.43.tgz: 2 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel-mix-6.0.43.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/terser/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-25858](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | detected in multiple dependencies | Transitive | N/A | &#10060; | | [CVE-2021-35065](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | glob-parent-5.1.2.tgz | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25858</summary> ### Vulnerable Libraries - <b>terser-4.8.0.tgz</b>, <b>terser-5.12.1.tgz</b></p> <p> ### <b>terser-4.8.0.tgz</b></p> <p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p> <p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-4.8.0.tgz">https://registry.npmjs.org/terser/-/terser-4.8.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/html-minifier-terser/node_modules/terser/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.43.tgz (Root Library) - html-loader-1.3.2.tgz - html-minifier-terser-5.1.1.tgz - :x: **terser-4.8.0.tgz** (Vulnerable Library) ### <b>terser-5.12.1.tgz</b></p> <p>JavaScript parser, mangler/compressor and beautifier toolkit for ES6+</p> <p>Library home page: <a href="https://registry.npmjs.org/terser/-/terser-5.12.1.tgz">https://registry.npmjs.org/terser/-/terser-5.12.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/terser/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.43.tgz (Root Library) - :x: **terser-5.12.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The package terser before 4.8.1, from 5.0.0 and before 5.14.2 are vulnerable to Regular Expression Denial of Service (ReDoS) due to insecure usage of regular expressions. <p>Publish Date: 2022-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25858>CVE-2022-25858</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25858</a></p> <p>Release Date: 2022-07-15</p> <p>Fix Resolution: terser - 4.8.1,5.14.2</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-35065</summary> ### Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.43.tgz (Root Library) - chokidar-3.5.3.tgz - :x: **glob-parent-5.1.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TheFloAnd/eventlisting-laravel/commit/4b81300cb80123e4a6d7adef1acdcc193a8070dc">4b81300cb80123e4a6d7adef1acdcc193a8070dc</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_process
laravel mix tgz vulnerabilities highest severity is autoclosed vulnerable library laravel mix tgz path to dependency file package json path to vulnerable library node modules terser package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high detected in multiple dependencies transitive n a high glob parent tgz transitive n a details cve vulnerable libraries terser tgz terser tgz terser tgz javascript parser mangler compressor and beautifier toolkit for library home page a href path to dependency file package json path to vulnerable library node modules html minifier terser node modules terser package json dependency hierarchy laravel mix tgz root library html loader tgz html minifier terser tgz x terser tgz vulnerable library terser tgz javascript parser mangler compressor and beautifier toolkit for library home page a href path to dependency file package json path to vulnerable library node modules terser package json dependency hierarchy laravel mix tgz root library x terser tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package terser before from and before are vulnerable to regular expression denial of service redos due to insecure usage of regular expressions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution terser step up your open source security game with mend cve vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy laravel mix tgz root library chokidar tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with mend
0
279,751
24,252,594,201
IssuesEvent
2022-09-27 15:12:20
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Test badge on TreeView
testplan-item
Refs: https://github.com/microsoft/vscode/issues/62783 - [x] anyOS @meganrogge - [x] anyOS @bpasero Complexity: 4 Roles: Developer [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23161780%0A%0A&assignees=alexr00) --- We have newly finalized API for badges on TreeViews: https://github.com/microsoft/vscode/blob/9caff0557b9ed17a6ccb44b2f18c1f6915f2eb85/src/vscode-dts/vscode.d.ts#L10484-L10488 This badge shows as a circle with a number on the view's view container. To test: - Read the inline documentation for the API and verify that it's clear. - Try out the API in a tree view. For example, in the tree-view-sample: https://github.com/microsoft/vscode-extension-samples/blob/31d3d4920b3fae52e03a185e434f51ee81965515/tree-view-sample/src/testView.ts#L6-L7 - Verify that you can set the badge number value. - Verify that you can update the badge number value. - Verify that you can cause the badge to hide by unsetting the badge number value.
1.0
Test badge on TreeView - Refs: https://github.com/microsoft/vscode/issues/62783 - [x] anyOS @meganrogge - [x] anyOS @bpasero Complexity: 4 Roles: Developer [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23161780%0A%0A&assignees=alexr00) --- We have newly finalized API for badges on TreeViews: https://github.com/microsoft/vscode/blob/9caff0557b9ed17a6ccb44b2f18c1f6915f2eb85/src/vscode-dts/vscode.d.ts#L10484-L10488 This badge shows as a circle with a number on the view's view container. To test: - Read the inline documentation for the API and verify that it's clear. - Try out the API in a tree view. For example, in the tree-view-sample: https://github.com/microsoft/vscode-extension-samples/blob/31d3d4920b3fae52e03a185e434f51ee81965515/tree-view-sample/src/testView.ts#L6-L7 - Verify that you can set the badge number value. - Verify that you can update the badge number value. - Verify that you can cause the badge to hide by unsetting the badge number value.
non_process
test badge on treeview refs anyos meganrogge anyos bpasero complexity roles developer we have newly finalized api for badges on treeviews this badge shows as a circle with a number on the view s view container to test read the inline documentation for the api and verify that it s clear try out the api in a tree view for example in the tree view sample verify that you can set the badge number value verify that you can update the badge number value verify that you can cause the badge to hide by unsetting the badge number value
0
16,093
20,262,151,763
IssuesEvent
2022-02-15 08:39:47
sillsdev/silnlp
https://api.github.com/repos/sillsdev/silnlp
opened
The source vocab for a multilingual parent model is not correctly transferred to the child model
bug pipeline 3: preprocess
The source vocab for a multilingual model will contain special target language tags. When a child model is finetuned from a multilingual model, it will need to add a new target language tag to the vocab. Because of this, the preprocess script will build a new sentencepiece model, because the child vocab contains tokens that do not exist in the parent vocab. In this case, we should preserve the parent vocab while still adding the new target language tag.
1.0
The source vocab for a multilingual parent model is not correctly transferred to the child model - The source vocab for a multilingual model will contain special target language tags. When a child model is finetuned from a multilingual model, it will need to add a new target language tag to the vocab. Because of this, the preprocess script will build a new sentencepiece model, because the child vocab contains tokens that do not exist in the parent vocab. In this case, we should preserve the parent vocab while still adding the new target language tag.
process
the source vocab for a multilingual parent model is not correctly transferred to the child model the source vocab for a multilingual model will contain special target language tags when a child model is finetuned from a multilingual model it will need to add a new target language tag to the vocab because of this the preprocess script will build a new sentencepiece model because the child vocab contains tokens that do not exist in the parent vocab in this case we should preserve the parent vocab while still adding the new target language tag
1
160,794
13,797,237,431
IssuesEvent
2020-10-09 21:22:51
rrousselGit/river_pod
https://api.github.com/repos/rrousselGit/river_pod
closed
Usage of watch. Documentation shows 2 different usage of watch
documentation
Documentation shows 2 different usage of watch: `watch(xxxProvider.state)` and `watch(xxxProvider).state` , is there any difference? clarify when to use which?
1.0
Usage of watch. Documentation shows 2 different usage of watch - Documentation shows 2 different usage of watch: `watch(xxxProvider.state)` and `watch(xxxProvider).state` , is there any difference? clarify when to use which?
non_process
usage of watch documentation shows different usage of watch documentation shows different usage of watch watch xxxprovider state and watch xxxprovider state is there any difference clarify when to use which
0
7,739
10,862,521,073
IssuesEvent
2019-11-14 13:28:38
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Buying/selling grin not possible through bisq?
in:altcoins in:trade-process was:dropped
Grin was added to bisq in https://github.com/bisq-network/bisq/pull/2217 but as far as I can tell it's not possible to (reliably, or maybe at all) buy/sell grin through bisq for two reasons: 1. grin cannot be transferred without off-chain interaction between accounts in question (see e.g. https://blog.blockcypher.com/mining-grin-on-grinmint-how-to-request-payouts-1593835181eb). Easiest way with bisq seems to be that buyer of grin uses ip address as their grin address in bisq so that seller can send grin by connecting grin wallet/node directly to buyer's wallet/node. This feels awkward and might also be security issue since everything else in bisq runs over tor. 2. currently it's not possible to prove that grin transfer has taken place: https://github.com/mimblewimble/grin/issues/2336 Was grin added prematurely to bisq?
1.0
Buying/selling grin not possible through bisq? - Grin was added to bisq in https://github.com/bisq-network/bisq/pull/2217 but as far as I can tell it's not possible to (reliably, or maybe at all) buy/sell grin through bisq for two reasons: 1. grin cannot be transferred without off-chain interaction between accounts in question (see e.g. https://blog.blockcypher.com/mining-grin-on-grinmint-how-to-request-payouts-1593835181eb). Easiest way with bisq seems to be that buyer of grin uses ip address as their grin address in bisq so that seller can send grin by connecting grin wallet/node directly to buyer's wallet/node. This feels awkward and might also be security issue since everything else in bisq runs over tor. 2. currently it's not possible to prove that grin transfer has taken place: https://github.com/mimblewimble/grin/issues/2336 Was grin added prematurely to bisq?
process
buying selling grin not possible through bisq grin was added to bisq in but as far as i can tell it s not possible to reliably or maybe at all buy sell grin through bisq for two reasons grin cannot be transferred without off chain interaction between accounts in question see e g easiest way with bisq seems to be that buyer of grin uses ip address as their grin address in bisq so that seller can send grin by connecting grin wallet node directly to buyer s wallet node this feels awkward and might also be security issue since everything else in bisq runs over tor currently it s not possible to prove that grin transfer has taken place was grin added prematurely to bisq
1
291,606
21,931,781,228
IssuesEvent
2022-05-23 10:22:00
typeorm/typeorm
https://api.github.com/repos/typeorm/typeorm
closed
connection-api falsely imports getEntityManager instead of getManager
documentation requires triage
<!-- Please follow the template. If you don't, your issue may be closed. Have a question? This is the TypeORM issue tracker - and not the right place for general support or questions. Instead, check the "Support" Documentation on the best places to ask questions! https://github.com/typeorm/typeorm/blob/master/docs/support.md --> ## Documentation Issue ### What was unclear or otherwise insufficient? In connection-api, there's a usage of getEntityManager, but it should be changed to getManager <!-- If relevant, Please be clear about the documentation file, as well as the location within the file. Link to the documentation in the repository. If the page does not exist, please be clear why a new documentation section is needed. --> ### Recommended Fix PR#8504 <!-- How should we fix this documentation issue? Should we add examples, clarify the language, or drop the page entirely? --> ### Additional Context <!-- Add any other context about the documentation issue here. --> ### Are you willing to resolve this issue by submitting a Pull Request? <!-- Put "✅" (:white_check_mark:) to one of these options, left "✖️" (:heavy_multiplication_x:) others: --> - ✅ Yes, I have the time, and I know how to start. - ✖️ Yes, I have the time, but I don't know how to start. I would need guidance. - ✖️ No, I don’t have the time, but I can support (using donations) development. - ✖️ No, I don’t have the time and I’m okay to wait for the community / maintainers to resolve this issue. <!-- Remember that first-time contributors are welcome! 🙌 👋 Have a great day and thank you for the documentation problem report! -->
1.0
connection-api falsely imports getEntityManager instead of getManager - <!-- Please follow the template. If you don't, your issue may be closed. Have a question? This is the TypeORM issue tracker - and not the right place for general support or questions. Instead, check the "Support" Documentation on the best places to ask questions! https://github.com/typeorm/typeorm/blob/master/docs/support.md --> ## Documentation Issue ### What was unclear or otherwise insufficient? In connection-api, there's a usage of getEntityManager, but it should be changed to getManager <!-- If relevant, Please be clear about the documentation file, as well as the location within the file. Link to the documentation in the repository. If the page does not exist, please be clear why a new documentation section is needed. --> ### Recommended Fix PR#8504 <!-- How should we fix this documentation issue? Should we add examples, clarify the language, or drop the page entirely? --> ### Additional Context <!-- Add any other context about the documentation issue here. --> ### Are you willing to resolve this issue by submitting a Pull Request? <!-- Put "✅" (:white_check_mark:) to one of these options, left "✖️" (:heavy_multiplication_x:) others: --> - ✅ Yes, I have the time, and I know how to start. - ✖️ Yes, I have the time, but I don't know how to start. I would need guidance. - ✖️ No, I don’t have the time, but I can support (using donations) development. - ✖️ No, I don’t have the time and I’m okay to wait for the community / maintainers to resolve this issue. <!-- Remember that first-time contributors are welcome! 🙌 👋 Have a great day and thank you for the documentation problem report! -->
non_process
connection api falsely imports getentitymanager instead of getmanager please follow the template if you don t your issue may be closed have a question this is the typeorm issue tracker and not the right place for general support or questions instead check the support documentation on the best places to ask questions documentation issue what was unclear or otherwise insufficient in connection api there s a usage of getentitymanager but it should be changed to getmanager if relevant please be clear about the documentation file as well as the location within the file link to the documentation in the repository if the page does not exist please be clear why a new documentation section is needed recommended fix pr how should we fix this documentation issue should we add examples clarify the language or drop the page entirely additional context add any other context about the documentation issue here are you willing to resolve this issue by submitting a pull request ✅ yes i have the time and i know how to start ✖️ yes i have the time but i don t know how to start i would need guidance ✖️ no i don’t have the time but i can support using donations development ✖️ no i don’t have the time and i’m okay to wait for the community maintainers to resolve this issue remember that first time contributors are welcome 🙌 👋 have a great day and thank you for the documentation problem report
0
90,661
26,162,176,809
IssuesEvent
2022-12-31 18:33:58
sandboxie-plus/Sandboxie
https://api.github.com/repos/sandboxie-plus/Sandboxie
closed
SandMan: Explorer context menu does not work directly after a clean installation on Windows 11
fixed in next build Issue reproduced Win 11 SbieDll
### Describe what you noticed and did With a clean reinstallation, the **Explorer context menu** (shell integration) usually only works after a restart of Windows. ### Clean VM (running **Windows 11**, 21H2, 64-bit): 1.) Download and install the latest version of Sandboxie-Plus (v1.3.4, 64-bit). 2.) Click through the Setup Wizard "Personally", leaving everything at default values. 3.) Wait until the entry "Open Sandboxed" is added to the Explorer context menu. 4.) Run e.g. "Microsoft Edge" in the DefaultBox via the Explorer context menu. 5.) The execution does not work, but the following error messages are displayed: ![sandboxie_context_w11_b](https://user-images.githubusercontent.com/33252157/189864088-ca26bf7a-f744-4ada-bb72-2197e9305f21.jpg) `Start.exe (1504): SBIE2112 Object is not accessible: \Sessions\1\BaseNamedObjects\SM0:1504:120:WilError_03, call CreateMutant (C0000022) access=001F0001 initialized=1` After restarting Windows, it then works as it should. ~~**Note: Update 1 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1259720113**~~ ~~**Note: Update 2 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1282416049**~~ **Note: Update 3 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1328778347** ### How often did you encounter it so far? With each clean reinstall of Sandboxie Plus. ### Affected program Not relevant ### Download link Not relevant ### Where is the program located? Not relevant to my request. ### Expected behavior The Explorer context menu should work correctly immediately (as under SbieCtrl) or the user should be prompted to restart. ### What is your Windows edition and version? Windows 11 Pro, 21H2, 64-bit ### In which Windows account you have this problem? User account with UAC protection set to Always notify. ### Please mention any installed security software Microsoft Defender Antivirus ### What version of Sandboxie are you running? 1.3.2-1.6.3 ### Is it a new installation of Sandboxie? I recently did a new clean installation. ### Is it a regression? _No response_ ### In which sandbox type you have this problem? Not relevant to my request. ### Can you reproduce this problem on an empty sandbox? Not relevant to my request. ### Did you previously enable some security policy settings outside Sandboxie? _No response_ ### Crash dump _No response_ ### Trace log _No response_ ### Sandboxie.ini configuration _No response_
1.0
SandMan: Explorer context menu does not work directly after a clean installation on Windows 11 - ### Describe what you noticed and did With a clean reinstallation, the **Explorer context menu** (shell integration) usually only works after a restart of Windows. ### Clean VM (running **Windows 11**, 21H2, 64-bit): 1.) Download and install the latest version of Sandboxie-Plus (v1.3.4, 64-bit). 2.) Click through the Setup Wizard "Personally", leaving everything at default values. 3.) Wait until the entry "Open Sandboxed" is added to the Explorer context menu. 4.) Run e.g. "Microsoft Edge" in the DefaultBox via the Explorer context menu. 5.) The execution does not work, but the following error messages are displayed: ![sandboxie_context_w11_b](https://user-images.githubusercontent.com/33252157/189864088-ca26bf7a-f744-4ada-bb72-2197e9305f21.jpg) `Start.exe (1504): SBIE2112 Object is not accessible: \Sessions\1\BaseNamedObjects\SM0:1504:120:WilError_03, call CreateMutant (C0000022) access=001F0001 initialized=1` After restarting Windows, it then works as it should. ~~**Note: Update 1 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1259720113**~~ ~~**Note: Update 2 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1282416049**~~ **Note: Update 3 https://github.com/sandboxie-plus/Sandboxie/issues/2284#issuecomment-1328778347** ### How often did you encounter it so far? With each clean reinstall of Sandboxie Plus. ### Affected program Not relevant ### Download link Not relevant ### Where is the program located? Not relevant to my request. ### Expected behavior The Explorer context menu should work correctly immediately (as under SbieCtrl) or the user should be prompted to restart. ### What is your Windows edition and version? Windows 11 Pro, 21H2, 64-bit ### In which Windows account you have this problem? User account with UAC protection set to Always notify. ### Please mention any installed security software Microsoft Defender Antivirus ### What version of Sandboxie are you running? 1.3.2-1.6.3 ### Is it a new installation of Sandboxie? I recently did a new clean installation. ### Is it a regression? _No response_ ### In which sandbox type you have this problem? Not relevant to my request. ### Can you reproduce this problem on an empty sandbox? Not relevant to my request. ### Did you previously enable some security policy settings outside Sandboxie? _No response_ ### Crash dump _No response_ ### Trace log _No response_ ### Sandboxie.ini configuration _No response_
non_process
sandman explorer context menu does not work directly after a clean installation on windows describe what you noticed and did with a clean reinstallation the explorer context menu shell integration usually only works after a restart of windows clean vm running windows bit download and install the latest version of sandboxie plus bit click through the setup wizard personally leaving everything at default values wait until the entry open sandboxed is added to the explorer context menu run e g microsoft edge in the defaultbox via the explorer context menu the execution does not work but the following error messages are displayed start exe object is not accessible sessions basenamedobjects wilerror call createmutant access initialized after restarting windows it then works as it should note update note update note update how often did you encounter it so far with each clean reinstall of sandboxie plus affected program not relevant download link not relevant where is the program located not relevant to my request expected behavior the explorer context menu should work correctly immediately as under sbiectrl or the user should be prompted to restart what is your windows edition and version windows pro bit in which windows account you have this problem user account with uac protection set to always notify please mention any installed security software microsoft defender antivirus what version of sandboxie are you running is it a new installation of sandboxie i recently did a new clean installation is it a regression no response in which sandbox type you have this problem not relevant to my request can you reproduce this problem on an empty sandbox not relevant to my request did you previously enable some security policy settings outside sandboxie no response crash dump no response trace log no response sandboxie ini configuration no response
0
15,158
18,909,998,106
IssuesEvent
2021-11-16 13:12:51
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Better error message if using `TEXT` or `BLOB` in MySQL @id/@index/@unique
process/candidate topic: indexes team/migrations topic: extendedIndexes
We now say this is not allowed. We should point to the correct preview feature and mention about using the `length` argument for indexes to work with TEXT or BLOB fields.
1.0
Better error message if using `TEXT` or `BLOB` in MySQL @id/@index/@unique - We now say this is not allowed. We should point to the correct preview feature and mention about using the `length` argument for indexes to work with TEXT or BLOB fields.
process
better error message if using text or blob in mysql id index unique we now say this is not allowed we should point to the correct preview feature and mention about using the length argument for indexes to work with text or blob fields
1
52,493
27,592,009,460
IssuesEvent
2023-03-09 01:33:07
keras-team/keras
https://api.github.com/repos/keras-team/keras
closed
Can't excute ConvNeXt with input_tensor
type:bug/performance stat:awaiting response from contributor stale
When keras.applications.ConvNeXt is executed with input_tensor, it is not executed. The cause seems to be a problem that occurs as it changes to the form of '[input_tensor]' when passing utils.layer_utils.get_source_inputs. (When using input_shape, there is no problem.) This problem was identified in versions tensorflow==2.10 and keras==2.10. Check please. the following error message: ``` --------------------------------------------------------------------------- ValueError Traceback (most recent call last) Input In [79], in <cell line: 2>() 1 x = tf.keras.layers.Input(shape = (512,512,3)) ----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = True, include_top = False, weights = None) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 562 @keras_export( 563 "keras.applications.convnext.ConvNeXtTiny", 564 "keras.applications.ConvNeXtTiny", (...) 575 classifier_activation="softmax", 576 ): --> 577 return ConvNeXt( 578 depths=MODEL_CONFIGS["tiny"]["depths"], 579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"], 580 drop_path_rate=0.0, 581 layer_scale_init_value=1e-6, 582 default_size=MODEL_CONFIGS["tiny"]["default_size"], 583 model_name=model_name, 584 include_top=include_top, 585 include_preprocessing=include_preprocessing, 586 weights=weights, 587 input_tensor=input_tensor, 588 input_shape=input_shape, 589 pooling=pooling, 590 classes=classes, 591 classifier_activation=classifier_activation, 592 ) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:461, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 459 num_channels = input_shape[channel_axis - 1] 460 if num_channels == 3: --> 461 x = PreStem(name=model_name)(x) 463 # Stem block. 464 stem = sequential.Sequential( 465 [ 466 layers.Conv2D( (...) 476 name=model_name + "_stem", 477 ) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:313, in PreStem.<locals>.apply(x) 312 def apply(x): --> 313 x = layers.Normalization( 314 mean=[0.485 * 255, 0.456 * 255, 0.406 * 255], 315 variance=[ 316 (0.229 * 255) ** 2, 317 (0.224 * 255) ** 2, 318 (0.225 * 255) ** 2, 319 ], 320 name=name + "_prestem_normalization", 321 )(x) 322 return x File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs) 67 filtered_tb = _process_traceback_frames(e.__traceback__) 68 # To get the full stack trace, call: 69 # `tf.debugging.disable_traceback_filtering()` ---> 70 raise e.with_traceback(filtered_tb) from None 71 finally: 72 del filtered_tb File /opt/anaconda3/lib/python3.9/site-packages/keras/layers/preprocessing/normalization.py:166, in Normalization.build(self, input_shape) 161 super().build(input_shape) 163 if isinstance(input_shape, (list, tuple)) and all( 164 isinstance(shape, tf.TensorShape) for shape in input_shape 165 ): --> 166 raise ValueError( 167 "Normalization only accepts a single input. If you are " 168 "passing a python list or tuple as a single input, " 169 "please convert to a numpy array or `tf.Tensor`." 170 ) 172 input_shape = tf.TensorShape(input_shape).as_list() 173 ndim = len(input_shape) ValueError: Normalization only accepts a single input. If you are passing a python list or tuple as a single input, please convert to a numpy array or `tf.Tensor`. ``` the following error message without include_preprocessing: ``` WARNING:tensorflow:Layers in a Sequential model should only have a single input tensor. Received: inputs=[<tf.Tensor 'Placeholder:0' shape=(None, 512, 512, 3) dtype=float32>]. Consider rewriting this model with the Functional API. --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Input In [81], in <cell line: 2>() 1 x = tf.keras.layers.Input(shape = (512,512,3)) ----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = False, include_top = False, weights = None) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 562 @keras_export( 563 "keras.applications.convnext.ConvNeXtTiny", 564 "keras.applications.ConvNeXtTiny", (...) 575 classifier_activation="softmax", 576 ): --> 577 return ConvNeXt( 578 depths=MODEL_CONFIGS["tiny"]["depths"], 579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"], 580 drop_path_rate=0.0, 581 layer_scale_init_value=1e-6, 582 default_size=MODEL_CONFIGS["tiny"]["default_size"], 583 model_name=model_name, 584 include_top=include_top, 585 include_preprocessing=include_preprocessing, 586 weights=weights, 587 input_tensor=input_tensor, 588 input_shape=input_shape, 589 pooling=pooling, 590 classes=classes, 591 classifier_activation=classifier_activation, 592 ) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:514, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 512 num_convnext_blocks = 4 513 for i in range(num_convnext_blocks): --> 514 x = downsample_layers[i](x) 515 for j in range(depths[i]): 516 x = ConvNeXtBlock( 517 projection_dim=projection_dims[i], 518 drop_path_rate=depth_drop_rates[cur + j], 519 layer_scale_init_value=layer_scale_init_value, 520 name=model_name + f"_stage_{i}_block_{j}", 521 )(x) File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs) 67 filtered_tb = _process_traceback_frames(e.__traceback__) 68 # To get the full stack trace, call: 69 # `tf.debugging.disable_traceback_filtering()` ---> 70 raise e.with_traceback(filtered_tb) from None 71 finally: 72 del filtered_tb File <string>:3, in raise_from(value, from_value) TypeError: Exception encountered when calling layer "convnext_tiny_stem" (type Sequential). Dimension value must be integer or None or have an __index__ method, got value 'TensorShape([None, 512, 512, 3])' with type '<class 'tensorflow.python.framework.tensor_shape.TensorShape'>' Call arguments received by layer "convnext_tiny_stem" (type Sequential): • inputs=['tf.Tensor(shape=(None, 512, 512, 3), dtype=float32)'] • training=False • mask=None ```
True
Can't excute ConvNeXt with input_tensor - When keras.applications.ConvNeXt is executed with input_tensor, it is not executed. The cause seems to be a problem that occurs as it changes to the form of '[input_tensor]' when passing utils.layer_utils.get_source_inputs. (When using input_shape, there is no problem.) This problem was identified in versions tensorflow==2.10 and keras==2.10. Check please. the following error message: ``` --------------------------------------------------------------------------- ValueError Traceback (most recent call last) Input In [79], in <cell line: 2>() 1 x = tf.keras.layers.Input(shape = (512,512,3)) ----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = True, include_top = False, weights = None) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 562 @keras_export( 563 "keras.applications.convnext.ConvNeXtTiny", 564 "keras.applications.ConvNeXtTiny", (...) 575 classifier_activation="softmax", 576 ): --> 577 return ConvNeXt( 578 depths=MODEL_CONFIGS["tiny"]["depths"], 579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"], 580 drop_path_rate=0.0, 581 layer_scale_init_value=1e-6, 582 default_size=MODEL_CONFIGS["tiny"]["default_size"], 583 model_name=model_name, 584 include_top=include_top, 585 include_preprocessing=include_preprocessing, 586 weights=weights, 587 input_tensor=input_tensor, 588 input_shape=input_shape, 589 pooling=pooling, 590 classes=classes, 591 classifier_activation=classifier_activation, 592 ) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:461, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 459 num_channels = input_shape[channel_axis - 1] 460 if num_channels == 3: --> 461 x = PreStem(name=model_name)(x) 463 # Stem block. 464 stem = sequential.Sequential( 465 [ 466 layers.Conv2D( (...) 476 name=model_name + "_stem", 477 ) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:313, in PreStem.<locals>.apply(x) 312 def apply(x): --> 313 x = layers.Normalization( 314 mean=[0.485 * 255, 0.456 * 255, 0.406 * 255], 315 variance=[ 316 (0.229 * 255) ** 2, 317 (0.224 * 255) ** 2, 318 (0.225 * 255) ** 2, 319 ], 320 name=name + "_prestem_normalization", 321 )(x) 322 return x File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs) 67 filtered_tb = _process_traceback_frames(e.__traceback__) 68 # To get the full stack trace, call: 69 # `tf.debugging.disable_traceback_filtering()` ---> 70 raise e.with_traceback(filtered_tb) from None 71 finally: 72 del filtered_tb File /opt/anaconda3/lib/python3.9/site-packages/keras/layers/preprocessing/normalization.py:166, in Normalization.build(self, input_shape) 161 super().build(input_shape) 163 if isinstance(input_shape, (list, tuple)) and all( 164 isinstance(shape, tf.TensorShape) for shape in input_shape 165 ): --> 166 raise ValueError( 167 "Normalization only accepts a single input. If you are " 168 "passing a python list or tuple as a single input, " 169 "please convert to a numpy array or `tf.Tensor`." 170 ) 172 input_shape = tf.TensorShape(input_shape).as_list() 173 ndim = len(input_shape) ValueError: Normalization only accepts a single input. If you are passing a python list or tuple as a single input, please convert to a numpy array or `tf.Tensor`. ``` the following error message without include_preprocessing: ``` WARNING:tensorflow:Layers in a Sequential model should only have a single input tensor. Received: inputs=[<tf.Tensor 'Placeholder:0' shape=(None, 512, 512, 3) dtype=float32>]. Consider rewriting this model with the Functional API. --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Input In [81], in <cell line: 2>() 1 x = tf.keras.layers.Input(shape = (512,512,3)) ----> 2 model = tf.keras.applications.ConvNeXtTiny(input_tensor = x, include_preprocessing = False, include_top = False, weights = None) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:577, in ConvNeXtTiny(model_name, include_top, include_preprocessing, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 562 @keras_export( 563 "keras.applications.convnext.ConvNeXtTiny", 564 "keras.applications.ConvNeXtTiny", (...) 575 classifier_activation="softmax", 576 ): --> 577 return ConvNeXt( 578 depths=MODEL_CONFIGS["tiny"]["depths"], 579 projection_dims=MODEL_CONFIGS["tiny"]["projection_dims"], 580 drop_path_rate=0.0, 581 layer_scale_init_value=1e-6, 582 default_size=MODEL_CONFIGS["tiny"]["default_size"], 583 model_name=model_name, 584 include_top=include_top, 585 include_preprocessing=include_preprocessing, 586 weights=weights, 587 input_tensor=input_tensor, 588 input_shape=input_shape, 589 pooling=pooling, 590 classes=classes, 591 classifier_activation=classifier_activation, 592 ) File /opt/anaconda3/lib/python3.9/site-packages/keras/applications/convnext.py:514, in ConvNeXt(depths, projection_dims, drop_path_rate, layer_scale_init_value, default_size, model_name, include_preprocessing, include_top, weights, input_tensor, input_shape, pooling, classes, classifier_activation) 512 num_convnext_blocks = 4 513 for i in range(num_convnext_blocks): --> 514 x = downsample_layers[i](x) 515 for j in range(depths[i]): 516 x = ConvNeXtBlock( 517 projection_dim=projection_dims[i], 518 drop_path_rate=depth_drop_rates[cur + j], 519 layer_scale_init_value=layer_scale_init_value, 520 name=model_name + f"_stage_{i}_block_{j}", 521 )(x) File /opt/anaconda3/lib/python3.9/site-packages/keras/utils/traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs) 67 filtered_tb = _process_traceback_frames(e.__traceback__) 68 # To get the full stack trace, call: 69 # `tf.debugging.disable_traceback_filtering()` ---> 70 raise e.with_traceback(filtered_tb) from None 71 finally: 72 del filtered_tb File <string>:3, in raise_from(value, from_value) TypeError: Exception encountered when calling layer "convnext_tiny_stem" (type Sequential). Dimension value must be integer or None or have an __index__ method, got value 'TensorShape([None, 512, 512, 3])' with type '<class 'tensorflow.python.framework.tensor_shape.TensorShape'>' Call arguments received by layer "convnext_tiny_stem" (type Sequential): • inputs=['tf.Tensor(shape=(None, 512, 512, 3), dtype=float32)'] • training=False • mask=None ```
non_process
can t excute convnext with input tensor when keras applications convnext is executed with input tensor it is not executed the cause seems to be a problem that occurs as it changes to the form of when passing utils layer utils get source inputs when using input shape there is no problem this problem was identified in versions tensorflow and keras check please the following error message valueerror traceback most recent call last input in in x tf keras layers input shape model tf keras applications convnexttiny input tensor x include preprocessing true include top false weights none file opt lib site packages keras applications convnext py in convnexttiny model name include top include preprocessing weights input tensor input shape pooling classes classifier activation keras export keras applications convnext convnexttiny keras applications convnexttiny classifier activation softmax return convnext depths model configs projection dims model configs drop path rate layer scale init value default size model configs model name model name include top include top include preprocessing include preprocessing weights weights input tensor input tensor input shape input shape pooling pooling classes classes classifier activation classifier activation file opt lib site packages keras applications convnext py in convnext depths projection dims drop path rate layer scale init value default size model name include preprocessing include top weights input tensor input shape pooling classes classifier activation num channels input shape if num channels x prestem name model name x stem block stem sequential sequential layers name model name stem file opt lib site packages keras applications convnext py in prestem apply x def apply x x layers normalization mean variance name name prestem normalization x return x file opt lib site packages keras utils traceback utils py in filter traceback error handler args kwargs filtered tb process traceback frames e traceback to get the full stack trace call tf debugging disable traceback filtering raise e with traceback filtered tb from none finally del filtered tb file opt lib site packages keras layers preprocessing normalization py in normalization build self input shape super build input shape if isinstance input shape list tuple and all isinstance shape tf tensorshape for shape in input shape raise valueerror normalization only accepts a single input if you are passing a python list or tuple as a single input please convert to a numpy array or tf tensor input shape tf tensorshape input shape as list ndim len input shape valueerror normalization only accepts a single input if you are passing a python list or tuple as a single input please convert to a numpy array or tf tensor the following error message without include preprocessing warning tensorflow layers in a sequential model should only have a single input tensor received inputs consider rewriting this model with the functional api typeerror traceback most recent call last input in in x tf keras layers input shape model tf keras applications convnexttiny input tensor x include preprocessing false include top false weights none file opt lib site packages keras applications convnext py in convnexttiny model name include top include preprocessing weights input tensor input shape pooling classes classifier activation keras export keras applications convnext convnexttiny keras applications convnexttiny classifier activation softmax return convnext depths model configs projection dims model configs drop path rate layer scale init value default size model configs model name model name include top include top include preprocessing include preprocessing weights weights input tensor input tensor input shape input shape pooling pooling classes classes classifier activation classifier activation file opt lib site packages keras applications convnext py in convnext depths projection dims drop path rate layer scale init value default size model name include preprocessing include top weights input tensor input shape pooling classes classifier activation num convnext blocks for i in range num convnext blocks x downsample layers x for j in range depths x convnextblock projection dim projection dims drop path rate depth drop rates layer scale init value layer scale init value name model name f stage i block j x file opt lib site packages keras utils traceback utils py in filter traceback error handler args kwargs filtered tb process traceback frames e traceback to get the full stack trace call tf debugging disable traceback filtering raise e with traceback filtered tb from none finally del filtered tb file in raise from value from value typeerror exception encountered when calling layer convnext tiny stem type sequential dimension value must be integer or none or have an index method got value tensorshape with type call arguments received by layer convnext tiny stem type sequential • inputs • training false • mask none
0
21,413
29,359,588,912
IssuesEvent
2023-05-28 00:36:15
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] Node.js Developer na Coodesh
SALVADOR BACK-END INFRAESTRUTURA FULL-STACK SCRUM BDD GIT TYPESCRIPT NODE.JS DOCKER DEVOPS REACT AWS REMOTO PROCESSOS INOVAÇÃO BACKEND GITHUB KANBAN CI CD SEGURANÇA GITFLOW UMA C QUALIDADE CLEAN XP TESTES AUTOMATIZADOS MICROSERVICES METODOLOGIAS ÁGEIS EXPRESS NEGÓCIOS MONITORAMENTO SRE PAAS Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>O Grupo fácil está buscando <strong><ins>Node Developer</ins></strong> para fazer parte de seu time!</p> <p></p> <p><strong>🏢 Quem somos: </strong></p> <p>O Grupo Fácil está focado em desenvolvimento de produtos na área da saúde, com o objetivo principal em atender os grandes nomes em operadoras de saúde do mercado brasileiro.</p> <p>Há mais de 27 anos dedicados na entrega de valor aos nossos clientes, focados na melhoria contínua e evolução natural no mercado de software brasileiro.</p> <p>Atualmente ingressamos em uma nova grande jornada de inovação, onde novos processos e produtos serão elaborados para entregar mais valor a nossos clientes, por isso, esperamos que você faça parte dessa jornada conosco.&nbsp;</p> <p></p> <p><strong>💻 Como é nosso time? </strong></p> <p>Trabalhamos de forma remota, usando as práticas de comunicação assíncrona e workflow para o desenvolvimento ágil, a principal ideia é trabalhar duro, focados nas metas e objetivos, porém sem perder a liberdade e a inspiração que nos faz acordar cedo todos os dias.</p> <p>Temos pessoas focadas no negócio, no frontend, outros no backend e até no perfil full-stack, esperamos de você que nos mostre onde se sente melhor. 😉&nbsp;</p> <p></p> <p><strong>O que esperamos de você?</strong></p> <ul> <li>Transparência (comunicação clara e eficiente sem ruído)</li> <li>Senso crítico com as demandas e processos</li> <li>Experiência com desenvolvimento de grandes projeto (escala, performance, qualidade e etc)</li> <li>Experiência na codificação de testes automatizados (unit test, BDD, Integration test)</li> <li>Experiência com Git/Github e GitFlow (já usou github actions?!)</li> <li>Capacidade de pedir ajuda e/ou ajudar em momentos de crise/impedimentos</li> <li>Capacidade de "se virar", "correr atrás" de algo desconhecido, assumir riscos (com parcimônia)</li> <li>Capacidade de ensinar, instruir e orientar pessoas menos experientes do projeto ou time (multiplique o conhecimento)</li> </ul> <p></p> <p><strong>Vamos falar de código? </strong></p> <ul> <li>Aqui usamos Typescript com S.O.L.I.D</li> <li>Usamos microservices, docker, AWS e tudo que há de bom 😁</li> <li>A arquitetura é desacoplada, testada com Hexagonal (Ports and Adapters)</li> <li>No frontend usamos React, Styled-Components, Vite, Storybook e etc</li> <li>No backend usamos Node.js, Express/Fastify, TypeORM, Oracle e etc</li> <li>E mais alguns detalhes que ficaremos felizes em contar durante nossa entrevista técnica.</li> <li>E ai, se sente confortável com isso?</li> </ul> <p></p> ## Grupo Fácil: <p>Ao longo de 27 anos de história, o Grupo Fácil se tornou referência nacional em sistemas, softwares e serviços para a gestão de negócios nas áreas financeira e de crédito, da saúde e no setor imobiliário.</p> <p>O Grupo Fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes.&nbsp;</p><a href='https://coodesh.com/empresas/grupo-facil'>Veja mais no site</a> ## Habilidades: - Node.js - Express.js - TypeORM - Oracle ## Local: 100% Remoto ## Diferenciais: - Conhecimento de Clean Architecture e/ou Hexagonal Architecture; - Conhecimentos/Experiência de automação e/ou infraestrutura como código (DevOps, CI/CD); - Conhecimento de metodologias ágeis (Kanban, Scrum, XP, Scrumban); - Conhecimento e/ou certificação em PaaS/Clouds (Preferencialmente AWS); - Fila, processamento assíncrono, tópicos, modelo pub/sub; - Sabe algo sobre monitoramento, autoscalling, stress-test, load-test, SRE?? ## Benefícios: - Convênio com farmácia; - Participação nos lucros; - Vale refeição; - Vale transporte; - Parcerias e convênios; - Programas de saúde e bem-estar. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Node.js Developer na Grupo Fácil](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Back-End
1.0
[Remoto] Node.js Developer na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>O Grupo fácil está buscando <strong><ins>Node Developer</ins></strong> para fazer parte de seu time!</p> <p></p> <p><strong>🏢 Quem somos: </strong></p> <p>O Grupo Fácil está focado em desenvolvimento de produtos na área da saúde, com o objetivo principal em atender os grandes nomes em operadoras de saúde do mercado brasileiro.</p> <p>Há mais de 27 anos dedicados na entrega de valor aos nossos clientes, focados na melhoria contínua e evolução natural no mercado de software brasileiro.</p> <p>Atualmente ingressamos em uma nova grande jornada de inovação, onde novos processos e produtos serão elaborados para entregar mais valor a nossos clientes, por isso, esperamos que você faça parte dessa jornada conosco.&nbsp;</p> <p></p> <p><strong>💻 Como é nosso time? </strong></p> <p>Trabalhamos de forma remota, usando as práticas de comunicação assíncrona e workflow para o desenvolvimento ágil, a principal ideia é trabalhar duro, focados nas metas e objetivos, porém sem perder a liberdade e a inspiração que nos faz acordar cedo todos os dias.</p> <p>Temos pessoas focadas no negócio, no frontend, outros no backend e até no perfil full-stack, esperamos de você que nos mostre onde se sente melhor. 😉&nbsp;</p> <p></p> <p><strong>O que esperamos de você?</strong></p> <ul> <li>Transparência (comunicação clara e eficiente sem ruído)</li> <li>Senso crítico com as demandas e processos</li> <li>Experiência com desenvolvimento de grandes projeto (escala, performance, qualidade e etc)</li> <li>Experiência na codificação de testes automatizados (unit test, BDD, Integration test)</li> <li>Experiência com Git/Github e GitFlow (já usou github actions?!)</li> <li>Capacidade de pedir ajuda e/ou ajudar em momentos de crise/impedimentos</li> <li>Capacidade de "se virar", "correr atrás" de algo desconhecido, assumir riscos (com parcimônia)</li> <li>Capacidade de ensinar, instruir e orientar pessoas menos experientes do projeto ou time (multiplique o conhecimento)</li> </ul> <p></p> <p><strong>Vamos falar de código? </strong></p> <ul> <li>Aqui usamos Typescript com S.O.L.I.D</li> <li>Usamos microservices, docker, AWS e tudo que há de bom 😁</li> <li>A arquitetura é desacoplada, testada com Hexagonal (Ports and Adapters)</li> <li>No frontend usamos React, Styled-Components, Vite, Storybook e etc</li> <li>No backend usamos Node.js, Express/Fastify, TypeORM, Oracle e etc</li> <li>E mais alguns detalhes que ficaremos felizes em contar durante nossa entrevista técnica.</li> <li>E ai, se sente confortável com isso?</li> </ul> <p></p> ## Grupo Fácil: <p>Ao longo de 27 anos de história, o Grupo Fácil se tornou referência nacional em sistemas, softwares e serviços para a gestão de negócios nas áreas financeira e de crédito, da saúde e no setor imobiliário.</p> <p>O Grupo Fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes.&nbsp;</p><a href='https://coodesh.com/empresas/grupo-facil'>Veja mais no site</a> ## Habilidades: - Node.js - Express.js - TypeORM - Oracle ## Local: 100% Remoto ## Diferenciais: - Conhecimento de Clean Architecture e/ou Hexagonal Architecture; - Conhecimentos/Experiência de automação e/ou infraestrutura como código (DevOps, CI/CD); - Conhecimento de metodologias ágeis (Kanban, Scrum, XP, Scrumban); - Conhecimento e/ou certificação em PaaS/Clouds (Preferencialmente AWS); - Fila, processamento assíncrono, tópicos, modelo pub/sub; - Sabe algo sobre monitoramento, autoscalling, stress-test, load-test, SRE?? ## Benefícios: - Convênio com farmácia; - Participação nos lucros; - Vale refeição; - Vale transporte; - Parcerias e convênios; - Programas de saúde e bem-estar. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Node.js Developer na Grupo Fácil](https://coodesh.com/vagas/nodejs-developer-plsr-131641269?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Back-End
process
node js developer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 o grupo fácil está buscando node developer para fazer parte de seu time 🏢 quem somos o grupo fácil está focado em desenvolvimento de produtos na área da saúde com o objetivo principal em atender os grandes nomes em operadoras de saúde do mercado brasileiro há mais de anos dedicados na entrega de valor aos nossos clientes focados na melhoria contínua e evolução natural no mercado de software brasileiro atualmente ingressamos em uma nova grande jornada de inovação onde novos processos e produtos serão elaborados para entregar mais valor a nossos clientes por isso esperamos que você faça parte dessa jornada conosco nbsp 💻 como é nosso time trabalhamos de forma remota usando as práticas de comunicação assíncrona e workflow para o desenvolvimento ágil a principal ideia é trabalhar duro focados nas metas e objetivos porém sem perder a liberdade e a inspiração que nos faz acordar cedo todos os dias temos pessoas focadas no negócio no frontend outros no backend e até no perfil full stack esperamos de você que nos mostre onde se sente melhor 😉 nbsp o que esperamos de você transparência comunicação clara e eficiente sem ruído senso crítico com as demandas e processos experiência com desenvolvimento de grandes projeto escala performance qualidade e etc experiência na codificação de testes automatizados unit test bdd integration test experiência com git github e gitflow já usou github actions capacidade de pedir ajuda e ou ajudar em momentos de crise impedimentos capacidade de se virar correr atrás de algo desconhecido assumir riscos com parcimônia capacidade de ensinar instruir e orientar pessoas menos experientes do projeto ou time multiplique o conhecimento vamos falar de código aqui usamos typescript com s o l i d usamos microservices docker aws e tudo que há de bom 😁 a arquitetura é desacoplada testada com hexagonal ports and adapters no frontend usamos react styled components vite storybook e etc no backend usamos node js express fastify typeorm oracle e etc e mais alguns detalhes que ficaremos felizes em contar durante nossa entrevista técnica e ai se sente confortável com isso grupo fácil ao longo de anos de história o grupo fácil se tornou referência nacional em sistemas softwares e serviços para a gestão de negócios nas áreas financeira e de crédito da saúde e no setor imobiliário o grupo fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes nbsp habilidades node js express js typeorm oracle local remoto diferenciais conhecimento de clean architecture e ou hexagonal architecture conhecimentos experiência de automação e ou infraestrutura como código devops ci cd conhecimento de metodologias ágeis kanban scrum xp scrumban conhecimento e ou certificação em paas clouds preferencialmente aws fila processamento assíncrono tópicos modelo pub sub sabe algo sobre monitoramento autoscalling stress test load test sre benefícios convênio com farmácia participação nos lucros vale refeição vale transporte parcerias e convênios programas de saúde e bem estar como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria back end
1
11,876
14,675,535,217
IssuesEvent
2020-12-30 17:50:40
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
gdal polygonize fails with "No module named '_gdal'. message
Bug Feedback Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> Running Raster -> Conversion -> Polygonize fails. The log contains the exception ModuleNotFoundError with message "No module named 'osgeo._gdal'" Handling this exception produces another with message "No module named '_gdal'. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> Load a raster. Use menu Raster -> Conversion -> Polygonize Choose raster as input layer and click Run **QGIS and OS versions** <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> The error appears in 3.10.13. It is not in 3.10.8. QGIS version | 3.10.13-A Coruña | QGIS code revision | 311f3bbade -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4 Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020 OS Version | Windows 10 (10.0) Active python plugins | plugin_reloader; QSWAT3; QSWAT3_64; QSWATPlus3_64; remotedebug; db_manager; MetaSearch; processing **Additional context** <!-- Add any other context about the problem here. -->
1.0
gdal polygonize fails with "No module named '_gdal'. message - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> Running Raster -> Conversion -> Polygonize fails. The log contains the exception ModuleNotFoundError with message "No module named 'osgeo._gdal'" Handling this exception produces another with message "No module named '_gdal'. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> Load a raster. Use menu Raster -> Conversion -> Polygonize Choose raster as input layer and click Run **QGIS and OS versions** <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> The error appears in 3.10.13. It is not in 3.10.8. QGIS version | 3.10.13-A Coruña | QGIS code revision | 311f3bbade -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4 Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020 OS Version | Windows 10 (10.0) Active python plugins | plugin_reloader; QSWAT3; QSWAT3_64; QSWATPlus3_64; remotedebug; db_manager; MetaSearch; processing **Additional context** <!-- Add any other context about the problem here. -->
process
gdal polygonize fails with no module named gdal message bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug running raster conversion polygonize fails the log contains the exception modulenotfounderror with message no module named osgeo gdal handling this exception produces another with message no module named gdal how to reproduce load a raster use menu raster conversion polygonize choose raster as input layer and click run qgis and os versions about click in the table ctrl a and then ctrl c finally paste here the error appears in it is not in qgis version a coruña qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins plugin reloader remotedebug db manager metasearch processing additional context
1
183,348
6,687,954,207
IssuesEvent
2017-10-08 08:18:02
CruCentralCoast/CruiOS
https://api.github.com/repos/CruCentralCoast/CruiOS
closed
About Section: Google Form Link is Editable
Priority: Minor Type: Bug
On the About section, we have a link to a Google Form for developer feedback. Instead of going to the link, the app opens a keyboard and allows you to type more letters and numbers onto the URL. I have no idea if this affected the data on the server, so if it did, the Google Form link will have to be changed back to what it was. The URL should act as a URL and open the form. iOS 11.0.2 Version 1.4 (20) iPhone 6
1.0
About Section: Google Form Link is Editable - On the About section, we have a link to a Google Form for developer feedback. Instead of going to the link, the app opens a keyboard and allows you to type more letters and numbers onto the URL. I have no idea if this affected the data on the server, so if it did, the Google Form link will have to be changed back to what it was. The URL should act as a URL and open the form. iOS 11.0.2 Version 1.4 (20) iPhone 6
non_process
about section google form link is editable on the about section we have a link to a google form for developer feedback instead of going to the link the app opens a keyboard and allows you to type more letters and numbers onto the url i have no idea if this affected the data on the server so if it did the google form link will have to be changed back to what it was the url should act as a url and open the form ios version iphone
0
319,630
27,389,839,343
IssuesEvent
2023-02-28 15:36:05
USEPA/haztrak
https://api.github.com/repos/USEPA/haztrak
opened
Quicker SIgn button and form
feature reactjs test
# 🚀 Feature Request Add a button to the manifest view to submit a Quicker Sign signature to e-Manifest. The Button should be disabled when... - The manifest is editable. the `readOnly` prop should be equal to `true`. - The user's RcraProfile.is_api_user field is `False` - The manifest status signifies that it is not time for the site, who the user is currently viewing as, to sign (awaiting signature of earlier handler, already been signed, etc) ## Additional to do items - This information is not currently stored in the Redux store (see new issue) - This information is not currently serialized to JSON when requested - We need to expand on this property to account for groups to allow admins to add non-API credential holders to use the RCRAInfo web services using another user's API credentials. <!-- Please provide a clear and concise description of what you'd like to see. Things like... 1. what is the expected behavior? 2. your opinion on what the default(s) should be? -->
1.0
Quicker SIgn button and form - # 🚀 Feature Request Add a button to the manifest view to submit a Quicker Sign signature to e-Manifest. The Button should be disabled when... - The manifest is editable. the `readOnly` prop should be equal to `true`. - The user's RcraProfile.is_api_user field is `False` - The manifest status signifies that it is not time for the site, who the user is currently viewing as, to sign (awaiting signature of earlier handler, already been signed, etc) ## Additional to do items - This information is not currently stored in the Redux store (see new issue) - This information is not currently serialized to JSON when requested - We need to expand on this property to account for groups to allow admins to add non-API credential holders to use the RCRAInfo web services using another user's API credentials. <!-- Please provide a clear and concise description of what you'd like to see. Things like... 1. what is the expected behavior? 2. your opinion on what the default(s) should be? -->
non_process
quicker sign button and form 🚀 feature request add a button to the manifest view to submit a quicker sign signature to e manifest the button should be disabled when the manifest is editable the readonly prop should be equal to true the user s rcraprofile is api user field is false the manifest status signifies that it is not time for the site who the user is currently viewing as to sign awaiting signature of earlier handler already been signed etc additional to do items this information is not currently stored in the redux store see new issue this information is not currently serialized to json when requested we need to expand on this property to account for groups to allow admins to add non api credential holders to use the rcrainfo web services using another user s api credentials please provide a clear and concise description of what you d like to see things like what is the expected behavior your opinion on what the default s should be
0
1,028
3,487,681,885
IssuesEvent
2016-01-02 05:28:21
spootTheLousy/saguaro
https://api.github.com/repos/spootTheLousy/saguaro
opened
Deletion class doesn't check for ghost bumping.
Administrative Post/text processing REVISIT
TO-DO: Prevent users from ghost bumping. Deletion timers maybe?
1.0
Deletion class doesn't check for ghost bumping. - TO-DO: Prevent users from ghost bumping. Deletion timers maybe?
process
deletion class doesn t check for ghost bumping to do prevent users from ghost bumping deletion timers maybe
1
15,981
20,188,186,289
IssuesEvent
2022-02-11 01:16:17
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Protect all public endpoints with appropriate controls
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Networking & Connectivity Endpoints
<a href="https://docs.microsoft.com/azure/architecture/framework/security/design-network-endpoints#web-application-firewalls-wafs">Protect all public endpoints with appropriate controls</a> <p><b>Why Consider This?</b></p> External application endpoints should be protected against common attack vectors, from Denial of Service (DoS) attacks like Slowloris to app-level exploits, to prevent potential application downtime due to malicious intent. Azure-native technologies such as Azure Firewall, Application Gateway/Azure Front Door, WAF, and DDoS Protection Standard Plan can be used to achieve requisite protection (Azure DDoS Protection). <p><b>Context</b></p> <p><b>Suggested Actions</b></p> <p><span>Protect all public endpoints with appropriate solutions, e.g. Azure Front Door, Application Gateway, Azure Firewall, Azure DDOS Protection or any 3rd party solution.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/firewall/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/firewall/overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/frontdoor/" target="_blank"><span>https://docs.microsoft.com/en-us/azure/frontdoor/</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/application-gateway/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/application-gateway/overview</span></a><span /></p>
1.0
Protect all public endpoints with appropriate controls - <a href="https://docs.microsoft.com/azure/architecture/framework/security/design-network-endpoints#web-application-firewalls-wafs">Protect all public endpoints with appropriate controls</a> <p><b>Why Consider This?</b></p> External application endpoints should be protected against common attack vectors, from Denial of Service (DoS) attacks like Slowloris to app-level exploits, to prevent potential application downtime due to malicious intent. Azure-native technologies such as Azure Firewall, Application Gateway/Azure Front Door, WAF, and DDoS Protection Standard Plan can be used to achieve requisite protection (Azure DDoS Protection). <p><b>Context</b></p> <p><b>Suggested Actions</b></p> <p><span>Protect all public endpoints with appropriate solutions, e.g. Azure Front Door, Application Gateway, Azure Firewall, Azure DDOS Protection or any 3rd party solution.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/firewall/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/firewall/overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/virtual-network/ddos-protection-overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/frontdoor/" target="_blank"><span>https://docs.microsoft.com/en-us/azure/frontdoor/</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/application-gateway/overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/application-gateway/overview</span></a><span /></p>
process
protect all public endpoints with appropriate controls why consider this external application endpoints should be protected against common attack vectors from denial of service dos attacks like slowloris to app level exploits to prevent potential application downtime due to malicious intent azure native technologies such as azure firewall application gateway azure front door waf and ddos protection standard plan can be used to achieve requisite protection azure ddos protection context suggested actions protect all public endpoints with appropriate solutions e g azure front door application gateway azure firewall azure ddos protection or any party solution learn more
1
2,908
5,891,937,182
IssuesEvent
2017-05-17 18:16:21
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process: closes before flushing all data
child_process
I use NodeJS v5.1.0 x64 on Windows 10. The following code ``` javascript var cp = require('child_process'); var stdout = ""; var stderr = ""; var process = cp.spawn("git", ["ls-remote", "--tags", "--heads", "git://github.com/jquery/jquery.git"]); process.stdout.on('data', function (data) { var text = data.toString(); stdout += text; }); process.stderr.on('data', function (data) { var text = data.toString(); stderr += text; }); process.on('error', function (error) { debugger; }); process.on('close', function (code) { debugger; }); process.on('exit', function (code) { debugger; }); ``` results in the following text in stdout and then the `exit` and `close` events are triggered. You can see that the response is incomplete. When I call the command in the Windows CMD directly, I get a complete response back. ``` ff619f3b978304700add92952485a5fb0a8fd80c refs/heads/1.12-stable 4fea389ad2b7515bc3a9081245979ab7be566b99 refs/heads/2.2-stable 83391859bda90f630bf02a5e04d82e9f1babeb1f refs/heads/compat 90347a78d69b15a515c462eb378adb7e6067f5dd refs/heads/killphp eaa3e9f0cfc68083556cf61195821d90e369f646 refs/heads/master bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0 bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0.1 dbee06de85859af59591813d3004e6695b8bb278 refs/tags/1.0.2 4e3da33c59fafe34e237585743e86e24ba81046e refs/tags/1.0.3 ed37ca66e5f79b4250f38cbd7a7f809ebfc15a20 refs/tags/1.0.4 73f55becc5e056ad28f6aec22ec014ffb3ab913b refs/tags/1.0a 2aa67026ebe6bea90fd137fc99b4c9422977e3f0 refs/tags/1.1 2550b1ea6a6eb6ba50d378a5dcfbc8146ca88bff refs/tags/1.1.1 92973279ea26c415f505d187276460115c668f17 refs/tags/1.1.2 fdd437800d983346c5936abedcf292c2a612b892 refs/tags/1.1.3 f70010b0ad13e3e4d6366287fd79636a9ceca1cd refs/tags/1.1.3.1 19f5cf0c893aa2ba2ffb00c00c7efc1db4383f3e refs/tags/1.1.3a 557ec9a92f40f97254e48240825fcb9be5a08e20 refs/tags/1.1.4 2c62b787b2cafb32af14edadc963b8ba3cb847b8 refs/tags/1.10.0 3a8fcde250da0db237396e9ab97fb099eb68c624 refs/tags/1.10.0-beta1 935eb19d87063ec30b717b3d45368d3c73cb591d refs/tags/1.10.1 16b079b164d62bd807c612806842a13bf9b04d17 refs/tags/1.10.2 91cd1587ed4a1d4f3834227e8eb3cc30cb385409 refs/tags/1.11.0 b44595824c3f23441eab2f34551587c5ae6fd31c refs/tags/1.11.0-beta1 02cb5be0bb5a49cda3a187d169d89d91874bbf1e refs/tags/1.11.0-beta2 5bb568cd61cb127ef4ebcc63bb48a37cd5af562d refs/tags/1.11.0-beta3 5c62f3cd019218e1b401ce2cf4cad644772e387f refs/tags/1.11.0-rc1 0d5ec2d8ac94a419ee47a39319c43ff9a7326b50 refs/tags/1.11.1 b6c4e50fb8f24dcf7bf38e053e68c4d0e7f059e8 refs/tags/1.11.1-beta1 baa21ff7a91ca165b63a5f900313da14c36fe255 refs/tags/1.11.1-rc1 b50c9a8d92f94f310138f4528e3a36b5336539a1 refs/tags/1.11.1-rc2 9690801db01709bfbff5f977d07fb7cc14472908 refs/tags/1.11.2 1472290917f17af05e98007136096784f9051fab refs/tags/1.11.3 8bcb016dfaa029a5e98cc3c25b9f999053217f36 refs/tags/1.1a d1390665b6e7d475d237cafcf8948440c9adfbf8 refs/tags/1.1b ce256312d518e16a6ee7df5f7f301b8bf9a1caf2 refs/tags/1.2 456da2d4a805012003df3c1b173148368d3536f7 refs/tags/1.2.1 6c81082a6b20880cbd5d35ec5c50d4270f3af2ce refs/tags/1.2.2 13a5fdb6a1571012c3e065a991fb4eff083057fb refs/tags/1.2.2b 3f839a69c5682419c921a1a2110289c8941556b0 refs/tags/1.2.2b2 d9a3c4fefaf84db7eba9b1f8204d1e134266b823 refs/tags/1.2.3 32ba00081ddd56c00655d0a7b7825ac947247cc6 refs/tags/1.2.3a 2c07690377522119891784ec11c89e87f0a12545 refs/tags/1.2.3b 2db27efd662693dd90fbf89ea1e4bbee7c9fb929 refs/tags/1.2.4 1ba2865b089fc06242004b7003b4f8dcfd92dc13 refs/tags/1.2.4a 44a6f5d0cb7f1124cd80e9b72d1fff9a46fe4c0b refs/tags/1.2.4b 456d8a62ccf3244c58ace2b315694ad54cfc05e3 refs/tags/1.2.5 ee931edfea95a0ed49aa4382b5c09cae6f7dafdb refs/tags/1.2.6 9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3 9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3.0 6df57d1b82c88ea85d2be081ee8f2362e1c362b7 refs/tags/1.3.1 8b79ea13656727e2449807a812dd0fd8c2783aad refs/tags/1.3.1rc1 13e9cde0efeadbbc5b3a375ee642bed394dd3a59 refs/tags/1.3.2 df1f656aae019bab1c0e84b41aca26714b4d0d5c refs/tags/1.3b1 c786f1097ee26095f2f7eb721a450c550775ba84 refs/tags/1.3b2 869afc46cd8229f41d82380f01830dd59a0084f7 refs/tags/1.3rc1 4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4 4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4.0 15d261b9673e153f9e9710da897de62c306afe53 refs/tags/1.4.1 9e4beb5ac258bfc834913c1781891ef49b2deb93 refs/tags/1.4.2 07458abcaf02b5e991223953081bb0dd5240e2c1 refs/tags/1.4.3 66a788ab837fd9b483740ee0b44c1e76f846aadd refs/tags/1.4.3rc1 3e5520fbdc7231b3f38e145020b40524c1e6654d refs/tags/1.4.3rc2 9b20df0375f37cc0997bd8ed3d1892a23967d3d5 refs/tags/1.4.4 c22623602ee3c9ff70df9e6f51132931d19bab04 refs/tags/1.4.4rc1 4d7f7453dca8e385cfb02089a8d5b2c47371bba3 refs/tags/1.4.4rc2 50708c070704c607a0ddc51679b93845c8fee5e1 refs/tags/1.4.4rc3 c90a1cb61b41a5686bc3764f144a156e734ed459 refs/tags/1.4a1 f745357baec82e0be59c651f70d1e489f237c398 refs/tags/1.4a2 857d8afd6917775b7cef6fd22e57dd9acec77f01 refs/tags/1.4rc1 675226eea749 ``` EDIT: The same happens on `NodeJS v4.2.2 x86`
1.0
child_process: closes before flushing all data - I use NodeJS v5.1.0 x64 on Windows 10. The following code ``` javascript var cp = require('child_process'); var stdout = ""; var stderr = ""; var process = cp.spawn("git", ["ls-remote", "--tags", "--heads", "git://github.com/jquery/jquery.git"]); process.stdout.on('data', function (data) { var text = data.toString(); stdout += text; }); process.stderr.on('data', function (data) { var text = data.toString(); stderr += text; }); process.on('error', function (error) { debugger; }); process.on('close', function (code) { debugger; }); process.on('exit', function (code) { debugger; }); ``` results in the following text in stdout and then the `exit` and `close` events are triggered. You can see that the response is incomplete. When I call the command in the Windows CMD directly, I get a complete response back. ``` ff619f3b978304700add92952485a5fb0a8fd80c refs/heads/1.12-stable 4fea389ad2b7515bc3a9081245979ab7be566b99 refs/heads/2.2-stable 83391859bda90f630bf02a5e04d82e9f1babeb1f refs/heads/compat 90347a78d69b15a515c462eb378adb7e6067f5dd refs/heads/killphp eaa3e9f0cfc68083556cf61195821d90e369f646 refs/heads/master bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0 bcc8a837055fe720579628d758b7034d6b520f2e refs/tags/1.0.1 dbee06de85859af59591813d3004e6695b8bb278 refs/tags/1.0.2 4e3da33c59fafe34e237585743e86e24ba81046e refs/tags/1.0.3 ed37ca66e5f79b4250f38cbd7a7f809ebfc15a20 refs/tags/1.0.4 73f55becc5e056ad28f6aec22ec014ffb3ab913b refs/tags/1.0a 2aa67026ebe6bea90fd137fc99b4c9422977e3f0 refs/tags/1.1 2550b1ea6a6eb6ba50d378a5dcfbc8146ca88bff refs/tags/1.1.1 92973279ea26c415f505d187276460115c668f17 refs/tags/1.1.2 fdd437800d983346c5936abedcf292c2a612b892 refs/tags/1.1.3 f70010b0ad13e3e4d6366287fd79636a9ceca1cd refs/tags/1.1.3.1 19f5cf0c893aa2ba2ffb00c00c7efc1db4383f3e refs/tags/1.1.3a 557ec9a92f40f97254e48240825fcb9be5a08e20 refs/tags/1.1.4 2c62b787b2cafb32af14edadc963b8ba3cb847b8 refs/tags/1.10.0 3a8fcde250da0db237396e9ab97fb099eb68c624 refs/tags/1.10.0-beta1 935eb19d87063ec30b717b3d45368d3c73cb591d refs/tags/1.10.1 16b079b164d62bd807c612806842a13bf9b04d17 refs/tags/1.10.2 91cd1587ed4a1d4f3834227e8eb3cc30cb385409 refs/tags/1.11.0 b44595824c3f23441eab2f34551587c5ae6fd31c refs/tags/1.11.0-beta1 02cb5be0bb5a49cda3a187d169d89d91874bbf1e refs/tags/1.11.0-beta2 5bb568cd61cb127ef4ebcc63bb48a37cd5af562d refs/tags/1.11.0-beta3 5c62f3cd019218e1b401ce2cf4cad644772e387f refs/tags/1.11.0-rc1 0d5ec2d8ac94a419ee47a39319c43ff9a7326b50 refs/tags/1.11.1 b6c4e50fb8f24dcf7bf38e053e68c4d0e7f059e8 refs/tags/1.11.1-beta1 baa21ff7a91ca165b63a5f900313da14c36fe255 refs/tags/1.11.1-rc1 b50c9a8d92f94f310138f4528e3a36b5336539a1 refs/tags/1.11.1-rc2 9690801db01709bfbff5f977d07fb7cc14472908 refs/tags/1.11.2 1472290917f17af05e98007136096784f9051fab refs/tags/1.11.3 8bcb016dfaa029a5e98cc3c25b9f999053217f36 refs/tags/1.1a d1390665b6e7d475d237cafcf8948440c9adfbf8 refs/tags/1.1b ce256312d518e16a6ee7df5f7f301b8bf9a1caf2 refs/tags/1.2 456da2d4a805012003df3c1b173148368d3536f7 refs/tags/1.2.1 6c81082a6b20880cbd5d35ec5c50d4270f3af2ce refs/tags/1.2.2 13a5fdb6a1571012c3e065a991fb4eff083057fb refs/tags/1.2.2b 3f839a69c5682419c921a1a2110289c8941556b0 refs/tags/1.2.2b2 d9a3c4fefaf84db7eba9b1f8204d1e134266b823 refs/tags/1.2.3 32ba00081ddd56c00655d0a7b7825ac947247cc6 refs/tags/1.2.3a 2c07690377522119891784ec11c89e87f0a12545 refs/tags/1.2.3b 2db27efd662693dd90fbf89ea1e4bbee7c9fb929 refs/tags/1.2.4 1ba2865b089fc06242004b7003b4f8dcfd92dc13 refs/tags/1.2.4a 44a6f5d0cb7f1124cd80e9b72d1fff9a46fe4c0b refs/tags/1.2.4b 456d8a62ccf3244c58ace2b315694ad54cfc05e3 refs/tags/1.2.5 ee931edfea95a0ed49aa4382b5c09cae6f7dafdb refs/tags/1.2.6 9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3 9dc7a3a7d20c0e75b927c956cba8de4dc30db077 refs/tags/1.3.0 6df57d1b82c88ea85d2be081ee8f2362e1c362b7 refs/tags/1.3.1 8b79ea13656727e2449807a812dd0fd8c2783aad refs/tags/1.3.1rc1 13e9cde0efeadbbc5b3a375ee642bed394dd3a59 refs/tags/1.3.2 df1f656aae019bab1c0e84b41aca26714b4d0d5c refs/tags/1.3b1 c786f1097ee26095f2f7eb721a450c550775ba84 refs/tags/1.3b2 869afc46cd8229f41d82380f01830dd59a0084f7 refs/tags/1.3rc1 4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4 4ce38155a8978c053f0b5938ef3e3f8729381e38 refs/tags/1.4.0 15d261b9673e153f9e9710da897de62c306afe53 refs/tags/1.4.1 9e4beb5ac258bfc834913c1781891ef49b2deb93 refs/tags/1.4.2 07458abcaf02b5e991223953081bb0dd5240e2c1 refs/tags/1.4.3 66a788ab837fd9b483740ee0b44c1e76f846aadd refs/tags/1.4.3rc1 3e5520fbdc7231b3f38e145020b40524c1e6654d refs/tags/1.4.3rc2 9b20df0375f37cc0997bd8ed3d1892a23967d3d5 refs/tags/1.4.4 c22623602ee3c9ff70df9e6f51132931d19bab04 refs/tags/1.4.4rc1 4d7f7453dca8e385cfb02089a8d5b2c47371bba3 refs/tags/1.4.4rc2 50708c070704c607a0ddc51679b93845c8fee5e1 refs/tags/1.4.4rc3 c90a1cb61b41a5686bc3764f144a156e734ed459 refs/tags/1.4a1 f745357baec82e0be59c651f70d1e489f237c398 refs/tags/1.4a2 857d8afd6917775b7cef6fd22e57dd9acec77f01 refs/tags/1.4rc1 675226eea749 ``` EDIT: The same happens on `NodeJS v4.2.2 x86`
process
child process closes before flushing all data i use nodejs on windows the following code javascript var cp require child process var stdout var stderr var process cp spawn git process stdout on data function data var text data tostring stdout text process stderr on data function data var text data tostring stderr text process on error function error debugger process on close function code debugger process on exit function code debugger results in the following text in stdout and then the exit and close events are triggered you can see that the response is incomplete when i call the command in the windows cmd directly i get a complete response back refs heads stable refs heads stable refs heads compat refs heads killphp refs heads master refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags refs tags edit the same happens on nodejs
1
738,390
25,556,503,700
IssuesEvent
2022-11-30 07:14:32
pc2ccs/pc2v9
https://api.github.com/repos/pc2ccs/pc2v9
closed
event-feed API endpoint returns incorrect ordinal field for "runs" testcases
bug high priority CLICS CCS Update
**Describe the issue**: Reported by DOMJudge: For `"runs" `test cases in the event-feed, the `"ordinal"` value should start at 1 not 0 as per the CLICS Specification. Currently, the first test case has an ordinal value of `"0"`. **To Reproduce**: Start up a contest that has some judged runs in it. Start the Event Feed. Connect to the event-feed endpoint and download the event feed. Note that the first testcase for each submission has an `"ordinal"` value of 0. Consequently, all following test cases are off by one for `"ordinal"` **Expected behavior**: The `"ordinal" `value of the first test case should start at 1, not 0. **Actual behavior**: The `"ordinal" `of the first test case starts at 0, and the rest of the test cases are off by one. **Environment**: Windows 11/Ubuntu 20.04 **Log Info**: **Screenshots**: if applicable, please add attachments showing screenshots to help explain the issue. **Additional context**: ``` {"type":"runs", "id":"pc2-4141", "op":"create", "data": {"id":"TestCase-2971170821107431354","judgement_id":"Run-25513900955186117","ordinal":0,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.607-05","contest_time":"00:25:52.371"}} {"type":"runs", "id":"pc2-4142", "op":"create", "data": {"id":"TestCase--1950512010137499320","judgement_id":"Run-25513900955186117","ordinal":1,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.656-05","contest_time":"00:25:52.420"}} {"type":"runs", "id":"pc2-4143", "op":"create", "data": {"id":"TestCase--7824024682353058102","judgement_id":"Run-25513900955186117","ordinal":2,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.726-05","contest_time":"00:25:52.490"}} {"type":"runs", "id":"pc2-4144", "op":"create", "data": {"id":"TestCase-5245915230241241075","judgement_id":"Run-25513900955186117","ordinal":3,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.798-05","contest_time":"00:25:52.562"}} ```
1.0
event-feed API endpoint returns incorrect ordinal field for "runs" testcases - **Describe the issue**: Reported by DOMJudge: For `"runs" `test cases in the event-feed, the `"ordinal"` value should start at 1 not 0 as per the CLICS Specification. Currently, the first test case has an ordinal value of `"0"`. **To Reproduce**: Start up a contest that has some judged runs in it. Start the Event Feed. Connect to the event-feed endpoint and download the event feed. Note that the first testcase for each submission has an `"ordinal"` value of 0. Consequently, all following test cases are off by one for `"ordinal"` **Expected behavior**: The `"ordinal" `value of the first test case should start at 1, not 0. **Actual behavior**: The `"ordinal" `of the first test case starts at 0, and the rest of the test cases are off by one. **Environment**: Windows 11/Ubuntu 20.04 **Log Info**: **Screenshots**: if applicable, please add attachments showing screenshots to help explain the issue. **Additional context**: ``` {"type":"runs", "id":"pc2-4141", "op":"create", "data": {"id":"TestCase-2971170821107431354","judgement_id":"Run-25513900955186117","ordinal":0,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.607-05","contest_time":"00:25:52.371"}} {"type":"runs", "id":"pc2-4142", "op":"create", "data": {"id":"TestCase--1950512010137499320","judgement_id":"Run-25513900955186117","ordinal":1,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.656-05","contest_time":"00:25:52.420"}} {"type":"runs", "id":"pc2-4143", "op":"create", "data": {"id":"TestCase--7824024682353058102","judgement_id":"Run-25513900955186117","ordinal":2,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.726-05","contest_time":"00:25:52.490"}} {"type":"runs", "id":"pc2-4144", "op":"create", "data": {"id":"TestCase-5245915230241241075","judgement_id":"Run-25513900955186117","ordinal":3,"judgement_type_id":"AC","time":"2022-11-08T07:47:25.798-05","contest_time":"00:25:52.562"}} ```
non_process
event feed api endpoint returns incorrect ordinal field for runs testcases describe the issue reported by domjudge for runs test cases in the event feed the ordinal value should start at not as per the clics specification currently the first test case has an ordinal value of to reproduce start up a contest that has some judged runs in it start the event feed connect to the event feed endpoint and download the event feed note that the first testcase for each submission has an ordinal value of consequently all following test cases are off by one for ordinal expected behavior the ordinal value of the first test case should start at not actual behavior the ordinal of the first test case starts at and the rest of the test cases are off by one environment windows ubuntu log info screenshots if applicable please add attachments showing screenshots to help explain the issue additional context type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time type runs id op create data id testcase judgement id run ordinal judgement type id ac time contest time
0
4,802
7,696,239,052
IssuesEvent
2018-05-18 14:44:54
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Refactor 'Code instrumentation' subsystem
!IMPORTANT! AREA: client AREA: server SYSTEM: resource processing TYPE: enhancement
We will use a native getter from property descriptor. It allows to drop server and client script processing, improves stack traces for client errors and etc. It will be done after updating mobile browsers `Android 6.0+` and `Safari 10+`.
1.0
Refactor 'Code instrumentation' subsystem - We will use a native getter from property descriptor. It allows to drop server and client script processing, improves stack traces for client errors and etc. It will be done after updating mobile browsers `Android 6.0+` and `Safari 10+`.
process
refactor code instrumentation subsystem we will use a native getter from property descriptor it allows to drop server and client script processing improves stack traces for client errors and etc it will be done after updating mobile browsers android and safari
1
137,035
12,743,816,124
IssuesEvent
2020-06-26 11:12:55
dask/dask-jobqueue
https://api.github.com/repos/dask/dask-jobqueue
closed
New Github design shows oldest release as latest?
documentation
The new side bar seems to get the order of releases wrong. It lists the v0.1 as the latest release: ![image](https://user-images.githubusercontent.com/5700886/85598583-03752500-b64c-11ea-8e6a-964edfb31d00.png) (Not sure if this is something that can be fixed in this repo or if this is a bug with Github.)
1.0
New Github design shows oldest release as latest? - The new side bar seems to get the order of releases wrong. It lists the v0.1 as the latest release: ![image](https://user-images.githubusercontent.com/5700886/85598583-03752500-b64c-11ea-8e6a-964edfb31d00.png) (Not sure if this is something that can be fixed in this repo or if this is a bug with Github.)
non_process
new github design shows oldest release as latest the new side bar seems to get the order of releases wrong it lists the as the latest release not sure if this is something that can be fixed in this repo or if this is a bug with github
0
69,694
17,792,168,391
IssuesEvent
2021-08-31 17:29:18
netlify/cli
https://api.github.com/repos/netlify/cli
closed
Support ODB in local builds and `netlify dev`
type: feature proj/odb-build
ODB should work in `netlify build` (local builds) and in `netlify dev`. This is a cover issue. More details should come as we implement the production behavior. See also: https://github.com/netlify/buildbot/issues/1645, https://github.com/netlify/buildbot/issues/1646, https://github.com/netlify/build/issues/3547, https://github.com/netlify/build/issues/3548, https://github.com/netlify/build/issues/3549, https://github.com/netlify/build/issues/3550
1.0
Support ODB in local builds and `netlify dev` - ODB should work in `netlify build` (local builds) and in `netlify dev`. This is a cover issue. More details should come as we implement the production behavior. See also: https://github.com/netlify/buildbot/issues/1645, https://github.com/netlify/buildbot/issues/1646, https://github.com/netlify/build/issues/3547, https://github.com/netlify/build/issues/3548, https://github.com/netlify/build/issues/3549, https://github.com/netlify/build/issues/3550
non_process
support odb in local builds and netlify dev odb should work in netlify build local builds and in netlify dev this is a cover issue more details should come as we implement the production behavior see also
0
286,864
31,773,270,358
IssuesEvent
2023-09-12 13:04:28
ssasy-auth/core
https://api.github.com/repos/ssasy-auth/core
closed
enhance challenge solution
question security
To solve a challenge, the current solution is to hash the challenge's nonce. An improvement would be to: - [ ] hash the entire challenge (without the solution property).
True
enhance challenge solution - To solve a challenge, the current solution is to hash the challenge's nonce. An improvement would be to: - [ ] hash the entire challenge (without the solution property).
non_process
enhance challenge solution to solve a challenge the current solution is to hash the challenge s nonce an improvement would be to hash the entire challenge without the solution property
0
153,366
24,118,325,343
IssuesEvent
2022-09-20 16:23:35
carbon-design-system/carbon-platform
https://api.github.com/repos/carbon-design-system/carbon-platform
opened
Resource card usage — When to use regular size versus small size?
role: design 🎨
🎨 Visual **Page:** https://next.carbondesignsystem.com/designing/get-started **Issue:** Resource card usage — When to use regular size versus small size? **Suggestion:** I was wondering if there are currently any rules around when to use each size of the Resource card component. On the `Designing > Get started` page, Step 2 (Access design tools) and Step 3 (tutorials) use different size resource cards. Was not sure if this was intentional or not. Looks like that maybe we are using the regular size resource card for the most important cards on the page (Design tools) and we use small ones for secondary link out content (Tutorials). Should they or shouldn't they coexist on the same page? I think we just need to outline some guidelines one when we use them and how we use them. ---- **Step 2 (Access design tools)** <img width="818" alt="Screen Shot 2022-09-19 at 4 17 31 PM" src="https://user-images.githubusercontent.com/43969356/191108663-05d4d15f-3799-4c49-8347-0ec8c037fb46.png"> **Step 3 (tutorials)** <img width="781" alt="Screen Shot 2022-09-19 at 4 17 42 PM" src="https://user-images.githubusercontent.com/43969356/191108688-6e3dc60f-2333-4fed-85a0-31ddd668ad45.png">
1.0
Resource card usage — When to use regular size versus small size? - 🎨 Visual **Page:** https://next.carbondesignsystem.com/designing/get-started **Issue:** Resource card usage — When to use regular size versus small size? **Suggestion:** I was wondering if there are currently any rules around when to use each size of the Resource card component. On the `Designing > Get started` page, Step 2 (Access design tools) and Step 3 (tutorials) use different size resource cards. Was not sure if this was intentional or not. Looks like that maybe we are using the regular size resource card for the most important cards on the page (Design tools) and we use small ones for secondary link out content (Tutorials). Should they or shouldn't they coexist on the same page? I think we just need to outline some guidelines one when we use them and how we use them. ---- **Step 2 (Access design tools)** <img width="818" alt="Screen Shot 2022-09-19 at 4 17 31 PM" src="https://user-images.githubusercontent.com/43969356/191108663-05d4d15f-3799-4c49-8347-0ec8c037fb46.png"> **Step 3 (tutorials)** <img width="781" alt="Screen Shot 2022-09-19 at 4 17 42 PM" src="https://user-images.githubusercontent.com/43969356/191108688-6e3dc60f-2333-4fed-85a0-31ddd668ad45.png">
non_process
resource card usage — when to use regular size versus small size 🎨 visual page issue resource card usage — when to use regular size versus small size suggestion i was wondering if there are currently any rules around when to use each size of the resource card component on the designing get started page step access design tools and step tutorials use different size resource cards was not sure if this was intentional or not looks like that maybe we are using the regular size resource card for the most important cards on the page design tools and we use small ones for secondary link out content tutorials should they or shouldn t they coexist on the same page i think we just need to outline some guidelines one when we use them and how we use them step access design tools img width alt screen shot at pm src step tutorials img width alt screen shot at pm src
0
268,858
8,415,266,657
IssuesEvent
2018-10-13 12:55:45
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
Symmetric translation fixes
Category: Translation Priority: High Status: In Progress Type: Bug Type: Support in progress
Initial list to look at: * Motor vehicle station * Shopping complex
1.0
Symmetric translation fixes - Initial list to look at: * Motor vehicle station * Shopping complex
non_process
symmetric translation fixes initial list to look at motor vehicle station shopping complex
0
16,514
2,909,794,534
IssuesEvent
2015-06-21 03:30:24
prettydiff/prettydiff
https://api.github.com/repos/prettydiff/prettydiff
closed
XML comparison sometimes gives erroneous errors with attributes in different order
Defect QA Underway
Prettydiff usually does a good job of ignoring differences just in the order of attributes. However, I've found some cases where it doesn't work. In the output, the attributes do not appear to be sorted on the offending lines. Here is an example gist containing two files to diff: https://gist.github.com/murrayju/5d99c370e06176044fa0 (I'm confident that they are equal, oXygen's diff tool agrees)
1.0
XML comparison sometimes gives erroneous errors with attributes in different order - Prettydiff usually does a good job of ignoring differences just in the order of attributes. However, I've found some cases where it doesn't work. In the output, the attributes do not appear to be sorted on the offending lines. Here is an example gist containing two files to diff: https://gist.github.com/murrayju/5d99c370e06176044fa0 (I'm confident that they are equal, oXygen's diff tool agrees)
non_process
xml comparison sometimes gives erroneous errors with attributes in different order prettydiff usually does a good job of ignoring differences just in the order of attributes however i ve found some cases where it doesn t work in the output the attributes do not appear to be sorted on the offending lines here is an example gist containing two files to diff i m confident that they are equal oxygen s diff tool agrees
0
181,923
21,664,468,081
IssuesEvent
2022-05-07 01:27:02
eldorplus/portfolio
https://api.github.com/repos/eldorplus/portfolio
closed
CVE-2018-11695 (High) detected in opennms-opennms-source-23.0.3-1 - autoclosed
security vulnerability
## CVE-2018-11695 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-23.0.3-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/eldorplus/portfolio/commits/d39c6030d1112cc864fbcffa04099b507e753f36">d39c6030d1112cc864fbcffa04099b507e753f36</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (13)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /portfolio/node_modules/node-sass/src/sass_context_wrapper.cpp - /portfolio/node_modules/node-sass/src/libsass/src/expand.cpp - /portfolio/node_modules/node-sass/src/libsass/src/parser.hpp - /portfolio/node_modules/node-sass/src/libsass/src/util.hpp - /portfolio/node_modules/node-sass/src/libsass/src/cssize.cpp - /portfolio/node_modules/node-gyp/gyp/pylib/gyp/MSVSUtil.py - /portfolio/node_modules/node-sass/src/libsass/src/functions.cpp - /portfolio/node_modules/node-sass/src/libsass/src/prelexer.cpp - /portfolio/node_modules/node-sass/src/callback_bridge.h - /portfolio/node_modules/node-sass/src/libsass/src/sass.cpp - /portfolio/node_modules/node-sass/src/sass_context_wrapper.h - /portfolio/node_modules/node-sass/src/libsass/src/eval.cpp - /portfolio/node_modules/node-sass/src/libsass/src/debugger.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.2. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695>CVE-2018-11695</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11695 (High) detected in opennms-opennms-source-23.0.3-1 - autoclosed - ## CVE-2018-11695 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-23.0.3-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/eldorplus/portfolio/commits/d39c6030d1112cc864fbcffa04099b507e753f36">d39c6030d1112cc864fbcffa04099b507e753f36</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (13)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /portfolio/node_modules/node-sass/src/sass_context_wrapper.cpp - /portfolio/node_modules/node-sass/src/libsass/src/expand.cpp - /portfolio/node_modules/node-sass/src/libsass/src/parser.hpp - /portfolio/node_modules/node-sass/src/libsass/src/util.hpp - /portfolio/node_modules/node-sass/src/libsass/src/cssize.cpp - /portfolio/node_modules/node-gyp/gyp/pylib/gyp/MSVSUtil.py - /portfolio/node_modules/node-sass/src/libsass/src/functions.cpp - /portfolio/node_modules/node-sass/src/libsass/src/prelexer.cpp - /portfolio/node_modules/node-sass/src/callback_bridge.h - /portfolio/node_modules/node-sass/src/libsass/src/sass.cpp - /portfolio/node_modules/node-sass/src/sass_context_wrapper.h - /portfolio/node_modules/node-sass/src/libsass/src/eval.cpp - /portfolio/node_modules/node-sass/src/libsass/src/debugger.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.2. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695>CVE-2018-11695</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in opennms opennms source autoclosed cve high severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries portfolio node modules node sass src sass context wrapper cpp portfolio node modules node sass src libsass src expand cpp portfolio node modules node sass src libsass src parser hpp portfolio node modules node sass src libsass src util hpp portfolio node modules node sass src libsass src cssize cpp portfolio node modules node gyp gyp pylib gyp msvsutil py portfolio node modules node sass src libsass src functions cpp portfolio node modules node sass src libsass src prelexer cpp portfolio node modules node sass src callback bridge h portfolio node modules node sass src libsass src sass cpp portfolio node modules node sass src sass context wrapper h portfolio node modules node sass src libsass src eval cpp portfolio node modules node sass src libsass src debugger hpp vulnerability details an issue was discovered in libsass through a null pointer dereference was found in the function sass expand operator which could be leveraged by an attacker to cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
0
804,267
29,481,765,826
IssuesEvent
2023-06-02 06:29:22
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.microsoft.com - design is broken
browser-chrome priority-critical
<!-- @browser: Chrome 114.0.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36 --> <!-- @reported_with: unknown --> **URL**: https://www.microsoft.com/en-us/edge?form=MA13FJ&exp=e00 **Browser / Version**: Chrome 114.0.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Design is broken **Description**: Items are misaligned **Steps to Reproduce**: Design is broken on the navbar. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/c99633d1-b0de-469e-9b32-3cb72764a337.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.microsoft.com - design is broken - <!-- @browser: Chrome 114.0.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36 --> <!-- @reported_with: unknown --> **URL**: https://www.microsoft.com/en-us/edge?form=MA13FJ&exp=e00 **Browser / Version**: Chrome 114.0.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Design is broken **Description**: Items are misaligned **Steps to Reproduce**: Design is broken on the navbar. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/c99633d1-b0de-469e-9b32-3cb72764a337.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
design is broken url browser version chrome operating system windows tested another browser yes edge problem type design is broken description items are misaligned steps to reproduce design is broken on the navbar view the screenshot img alt screenshot src browser configuration none from with ❤️
0
16,592
21,641,748,591
IssuesEvent
2022-05-05 19:34:06
MagicalAlexey/Keepers-of-the-Stones
https://api.github.com/repos/MagicalAlexey/Keepers-of-the-Stones
opened
Adding a full-fledged level system
new feature In process
There are plans to add a full-fledged pumping system of stones in order to improve the combat performance of stones. Tasks: - [ ] Add GUI skills and levels - [ ] Add an Element Coin - [ ] Add a way to get a coin - [ ] Add additional abilities
1.0
Adding a full-fledged level system - There are plans to add a full-fledged pumping system of stones in order to improve the combat performance of stones. Tasks: - [ ] Add GUI skills and levels - [ ] Add an Element Coin - [ ] Add a way to get a coin - [ ] Add additional abilities
process
adding a full fledged level system there are plans to add a full fledged pumping system of stones in order to improve the combat performance of stones tasks add gui skills and levels add an element coin add a way to get a coin add additional abilities
1
404,828
11,863,498,011
IssuesEvent
2020-03-25 19:50:20
near/near-api-js
https://api.github.com/repos/near/near-api-js
opened
Expose `genesis_config` and `genesis_records` RPC API
Priority 3 enhancement good first issue
We have recently added (nearprotocol/nearcore#2109) new RPC API methods: genesis_config and genesis_records (with `EXPERIMENTAL_` prefix). The API documentation is currently defined in the PR description mentioned above.
1.0
Expose `genesis_config` and `genesis_records` RPC API - We have recently added (nearprotocol/nearcore#2109) new RPC API methods: genesis_config and genesis_records (with `EXPERIMENTAL_` prefix). The API documentation is currently defined in the PR description mentioned above.
non_process
expose genesis config and genesis records rpc api we have recently added nearprotocol nearcore new rpc api methods genesis config and genesis records with experimental prefix the api documentation is currently defined in the pr description mentioned above
0
53,886
6,772,631,440
IssuesEvent
2017-10-27 00:07:19
HabitRPG/habitica
https://api.github.com/repos/HabitRPG/habitica
opened
Glasses and animal ears are not working
hacktoberfest help wanted POST-REDESIGN priority: important sections: Avatar/User Modal
Glasses and animal ears are not working. This may be happening because the Edit Avatar modal is equipping them to Battle Gear even when the user has Costume enabled.
1.0
Glasses and animal ears are not working - Glasses and animal ears are not working. This may be happening because the Edit Avatar modal is equipping them to Battle Gear even when the user has Costume enabled.
non_process
glasses and animal ears are not working glasses and animal ears are not working this may be happening because the edit avatar modal is equipping them to battle gear even when the user has costume enabled
0
21,096
28,045,730,679
IssuesEvent
2023-03-28 22:39:20
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
closed
Make ObjectMeta Non Exhaustive
enhancement development-process
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] (This section helps Arrow developers understand the context and *why* for this feature, in addition to the *what*) --> Currently the fields on `ObjectMeta` are public, making a breaking change to add additional metadata, such as in https://github.com/apache/arrow-rs/issues/2240 **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> I would like to make `ObjectMeta` [non_exhaustive](https://doc.rust-lang.org/reference/attributes/type_system.html#the-non_exhaustive-attribute) to allow adding new fields without requiring a breaking change. This will require adding a constructor to `ObjectMeta` to allow it to be safely constructed outside of the crate. **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
1.0
Make ObjectMeta Non Exhaustive - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] (This section helps Arrow developers understand the context and *why* for this feature, in addition to the *what*) --> Currently the fields on `ObjectMeta` are public, making a breaking change to add additional metadata, such as in https://github.com/apache/arrow-rs/issues/2240 **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> I would like to make `ObjectMeta` [non_exhaustive](https://doc.rust-lang.org/reference/attributes/type_system.html#the-non_exhaustive-attribute) to allow adding new fields without requiring a breaking change. This will require adding a constructor to `ObjectMeta` to allow it to be safely constructed outside of the crate. **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
process
make objectmeta non exhaustive is your feature request related to a problem or challenge please describe what you are trying to do a clear and concise description of what the problem is ex i m always frustrated when this section helps arrow developers understand the context and why for this feature in addition to the what currently the fields on objectmeta are public making a breaking change to add additional metadata such as in describe the solution you d like a clear and concise description of what you want to happen i would like to make objectmeta to allow adding new fields without requiring a breaking change this will require adding a constructor to objectmeta to allow it to be safely constructed outside of the crate describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
14,521
17,619,035,590
IssuesEvent
2021-08-18 13:22:00
linked-art/linked.art
https://api.github.com/repos/linked-art/linked.art
closed
Should we version the spec?
model defer process
I think semantic versioning of this would be really useful at some point. (as a 0.1 version, so we can make breaking changes as needed, but it would be helpful to indicate that this is in flux to people, and also to indicate when it _stops_ being in flux)
1.0
Should we version the spec? - I think semantic versioning of this would be really useful at some point. (as a 0.1 version, so we can make breaking changes as needed, but it would be helpful to indicate that this is in flux to people, and also to indicate when it _stops_ being in flux)
process
should we version the spec i think semantic versioning of this would be really useful at some point as a version so we can make breaking changes as needed but it would be helpful to indicate that this is in flux to people and also to indicate when it stops being in flux
1
37,882
15,391,048,812
IssuesEvent
2021-03-03 14:08:39
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[App Arch] Can't set Lucene Query String options through Advanced Settings
:AppServices/fix-it-week Feature:Query Bar Feature:Search Team:AppServices bug impact:high loe:hours triaged
If you try to add any query string options other than `analyze_wildcard` through Advanced Settings, you will get an error like ``` [validation [query:queryString:options].lenient]: definition for this key is missing ``` It appears that the schema for that setting only has the `analyze_wildcard` as a valid key and not any of the other valid query string options. https://github.com/elastic/kibana/blob/master/src/plugins/data/server/ui_settings.ts#L106
2.0
[App Arch] Can't set Lucene Query String options through Advanced Settings - If you try to add any query string options other than `analyze_wildcard` through Advanced Settings, you will get an error like ``` [validation [query:queryString:options].lenient]: definition for this key is missing ``` It appears that the schema for that setting only has the `analyze_wildcard` as a valid key and not any of the other valid query string options. https://github.com/elastic/kibana/blob/master/src/plugins/data/server/ui_settings.ts#L106
non_process
can t set lucene query string options through advanced settings if you try to add any query string options other than analyze wildcard through advanced settings you will get an error like lenient definition for this key is missing it appears that the schema for that setting only has the analyze wildcard as a valid key and not any of the other valid query string options
0
188,322
22,046,298,749
IssuesEvent
2022-05-30 02:22:15
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
closed
CVE-2016-10906 (High) detected in fedorav4.6 - autoclosed
security vulnerability
## CVE-2016-10906 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fedorav4.6</b></p></summary> <p> <p>Fedora kernel git tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git>https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in drivers/net/ethernet/arc/emac_main.c in the Linux kernel before 4.5. A use-after-free is caused by a race condition between the functions arc_emac_tx and arc_emac_tx_clean. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10906>CVE-2016-10906</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v4.5-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-10906 (High) detected in fedorav4.6 - autoclosed - ## CVE-2016-10906 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fedorav4.6</b></p></summary> <p> <p>Fedora kernel git tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git>https://git.kernel.org/pub/scm/linux/kernel/git/jwboyer/fedora.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/arc/emac_main.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in drivers/net/ethernet/arc/emac_main.c in the Linux kernel before 4.5. A use-after-free is caused by a race condition between the functions arc_emac_tx and arc_emac_tx_clean. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10906>CVE-2016-10906</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10906</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v4.5-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in autoclosed cve high severity vulnerability vulnerable library fedora kernel git tree library home page a href found in head commit a href found in base branch master vulnerable source files drivers net ethernet arc emac main c drivers net ethernet arc emac main c vulnerability details an issue was discovered in drivers net ethernet arc emac main c in the linux kernel before a use after free is caused by a race condition between the functions arc emac tx and arc emac tx clean publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
101,639
4,127,679,637
IssuesEvent
2016-06-10 00:11:38
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
Refactor SELinux related volume code in kubelet
area/storage priority/P2 team/cluster
Refactor SELinux related volume code in kubelet. For example the the code depends on being able to mutate the pod object.
1.0
Refactor SELinux related volume code in kubelet - Refactor SELinux related volume code in kubelet. For example the the code depends on being able to mutate the pod object.
non_process
refactor selinux related volume code in kubelet refactor selinux related volume code in kubelet for example the the code depends on being able to mutate the pod object
0
42,700
22,780,339,686
IssuesEvent
2022-07-08 18:57:10
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
opened
[Perf]: Move Form widget's Child widgets logic to derived properties
Performance Form Widget Task List Widget UI Performance
### Is there an existing issue for this? - [X] I have searched the existing issues ### SubTasks Form Widget uses childWidgets to get values to check each child's value, isValid, isDirty on the main Thread. This should Ideally be done in worker thread through derived Properties. This also helps us avoid Form widget from needing childWidgets which might also cause it to Re-render. This can also be tried for List Widget.
True
[Perf]: Move Form widget's Child widgets logic to derived properties - ### Is there an existing issue for this? - [X] I have searched the existing issues ### SubTasks Form Widget uses childWidgets to get values to check each child's value, isValid, isDirty on the main Thread. This should Ideally be done in worker thread through derived Properties. This also helps us avoid Form widget from needing childWidgets which might also cause it to Re-render. This can also be tried for List Widget.
non_process
move form widget s child widgets logic to derived properties is there an existing issue for this i have searched the existing issues subtasks form widget uses childwidgets to get values to check each child s value isvalid isdirty on the main thread this should ideally be done in worker thread through derived properties this also helps us avoid form widget from needing childwidgets which might also cause it to re render this can also be tried for list widget
0
17,459
23,282,747,984
IssuesEvent
2022-08-05 13:39:38
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag')
bug/1-unconfirmed kind/bug process/candidate tech/typescript topic: transaction team/client
### Bug description I try use prisma transaction and have error: ``` <project-home>/src/database/client/runtime/index.js:23284 const hasPreview = engine._hasPreviewFlag("tracing"); ^ TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag') at getTracingConfig (<project-home>/src/database/client/runtime/index.js:23284:29) at Object.$transaction [as transaction] (<project-home>/src/database/client/runtime/index.js:29563:29) ``` ### How to reproduce Simple code in my handler: ```js await prisma.$transaction([ prisma.passwordOfAdministrator.upsert({ where: { administratorId: confirmation.administratorId, }, create: { administratorId: confirmation.administratorId, hash: password.hash, salt: password.salt, createdAt, updatedAt, }, update: { hash: password.hash, salt: password.salt, updatedAt, }, }), prisma.recoveryPasswordConfirmation.update({ where: { id: confirmation.id, }, data: { ip, agent, confirmAt, }, }), ]) ``` ### Expected behavior Execute `prisma.$transaction` without error: ``` TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag') ``` ### Prisma information ``` datasource db { provider = "postgresql" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" output = "./client" engineType = "binary" } ``` or ``` datasource db { provider = "postgresql" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" output = "./client" engineType = "library" } ``` ### Environment & setup - OS: [Mac OS] - Database: [PostgreSQL] - Node.js version: [v16.15.0] ### Prisma Version ``` prisma : 4.1.1 @prisma/client : 4.1.1 Current platform : darwin Query Engine (Node-API) : libquery-engine 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node) Migration Engine : migration-engine-cli 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/migration-engine-darwin) Introspection Engine : introspection-core 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/introspection-engine-darwin) Format Binary : prisma-fmt 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/prisma-fmt-darwin) Default Engines Hash : 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 Studio : 0.469.0 ```
1.0
TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag') - ### Bug description I try use prisma transaction and have error: ``` <project-home>/src/database/client/runtime/index.js:23284 const hasPreview = engine._hasPreviewFlag("tracing"); ^ TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag') at getTracingConfig (<project-home>/src/database/client/runtime/index.js:23284:29) at Object.$transaction [as transaction] (<project-home>/src/database/client/runtime/index.js:29563:29) ``` ### How to reproduce Simple code in my handler: ```js await prisma.$transaction([ prisma.passwordOfAdministrator.upsert({ where: { administratorId: confirmation.administratorId, }, create: { administratorId: confirmation.administratorId, hash: password.hash, salt: password.salt, createdAt, updatedAt, }, update: { hash: password.hash, salt: password.salt, updatedAt, }, }), prisma.recoveryPasswordConfirmation.update({ where: { id: confirmation.id, }, data: { ip, agent, confirmAt, }, }), ]) ``` ### Expected behavior Execute `prisma.$transaction` without error: ``` TypeError: Cannot read properties of undefined (reading '_hasPreviewFlag') ``` ### Prisma information ``` datasource db { provider = "postgresql" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" output = "./client" engineType = "binary" } ``` or ``` datasource db { provider = "postgresql" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" output = "./client" engineType = "library" } ``` ### Environment & setup - OS: [Mac OS] - Database: [PostgreSQL] - Node.js version: [v16.15.0] ### Prisma Version ``` prisma : 4.1.1 @prisma/client : 4.1.1 Current platform : darwin Query Engine (Node-API) : libquery-engine 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node) Migration Engine : migration-engine-cli 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/migration-engine-darwin) Introspection Engine : introspection-core 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/introspection-engine-darwin) Format Binary : prisma-fmt 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 (at node_modules/@prisma/engines/prisma-fmt-darwin) Default Engines Hash : 8d8414deb360336e4698a65aa45a1fbaf1ce13d8 Studio : 0.469.0 ```
process
typeerror cannot read properties of undefined reading haspreviewflag bug description i try use prisma transaction and have error src database client runtime index js const haspreview engine haspreviewflag tracing typeerror cannot read properties of undefined reading haspreviewflag at gettracingconfig src database client runtime index js at object transaction src database client runtime index js how to reproduce simple code in my handler js await prisma transaction prisma passwordofadministrator upsert where administratorid confirmation administratorid create administratorid confirmation administratorid hash password hash salt password salt createdat updatedat update hash password hash salt password salt updatedat prisma recoverypasswordconfirmation update where id confirmation id data ip agent confirmat expected behavior execute prisma transaction without error typeerror cannot read properties of undefined reading haspreviewflag prisma information datasource db provider postgresql url env database url generator client provider prisma client js output client enginetype binary or datasource db provider postgresql url env database url generator client provider prisma client js output client enginetype library environment setup os database node js version prisma version prisma prisma client current platform darwin query engine node api libquery engine at node modules prisma engines libquery engine darwin dylib node migration engine migration engine cli at node modules prisma engines migration engine darwin introspection engine introspection core at node modules prisma engines introspection engine darwin format binary prisma fmt at node modules prisma engines prisma fmt darwin default engines hash studio
1
435,263
12,533,480,357
IssuesEvent
2020-06-04 17:40:43
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
Auth source rules are disabled by default
Priority: High Type: Bug
**Describe the bug** The auth source rules are shown as disabled by default. This should be the opposite **To Reproduce** 1. Create an AD source with an auth rule 2. The auth rule shows disabled by default
1.0
Auth source rules are disabled by default - **Describe the bug** The auth source rules are shown as disabled by default. This should be the opposite **To Reproduce** 1. Create an AD source with an auth rule 2. The auth rule shows disabled by default
non_process
auth source rules are disabled by default describe the bug the auth source rules are shown as disabled by default this should be the opposite to reproduce create an ad source with an auth rule the auth rule shows disabled by default
0
12,808
3,653,103,899
IssuesEvent
2016-02-17 05:01:59
Esri/angular-esri-map
https://api.github.com/repos/Esri/angular-esri-map
closed
Update scene view example pages to show warning for unsupported (mobile) devices
documentation enhancement
depends on #239 See this sample page for the pattern: https://developers.arcgis.com/javascript/beta/sample-code/scene-webgl-support/index.html
1.0
Update scene view example pages to show warning for unsupported (mobile) devices - depends on #239 See this sample page for the pattern: https://developers.arcgis.com/javascript/beta/sample-code/scene-webgl-support/index.html
non_process
update scene view example pages to show warning for unsupported mobile devices depends on see this sample page for the pattern
0
101,745
11,255,951,472
IssuesEvent
2020-01-12 13:08:33
ManoukK/NIBUD_datavisualisatie
https://api.github.com/repos/ManoukK/NIBUD_datavisualisatie
opened
README maken
documentation
readme make en afmaken Hierin moet ook staan hoe het werkt met de dataset van het HIBUD omdat die niet online kan
1.0
README maken - readme make en afmaken Hierin moet ook staan hoe het werkt met de dataset van het HIBUD omdat die niet online kan
non_process
readme maken readme make en afmaken hierin moet ook staan hoe het werkt met de dataset van het hibud omdat die niet online kan
0
186,939
15,088,052,655
IssuesEvent
2021-02-05 23:36:43
dapr/php-sdk
https://api.github.com/repos/dapr/php-sdk
closed
Add an editorconfig
documentation enhancement
Formatting is kind-of spotty and inconsistent. It'd be good to have some consistency and have formatting be enforced in PRs.
1.0
Add an editorconfig - Formatting is kind-of spotty and inconsistent. It'd be good to have some consistency and have formatting be enforced in PRs.
non_process
add an editorconfig formatting is kind of spotty and inconsistent it d be good to have some consistency and have formatting be enforced in prs
0
20,364
27,021,106,449
IssuesEvent
2023-02-11 02:25:13
openxla/stablehlo
https://api.github.com/repos/openxla/stablehlo
opened
Populate conformance suite with "interesting" test cases
Process
Our current methodology for testing the reference implementation involves writing a minimal amount of tests (https://github.com/openxla/stablehlo/blob/main/docs/reference_checklist.md), but for a conformance suite we'll likely want to write "interesting" tests that cover potentially surprising behavior, even if strictly speaking these tests don't increase coverage of the codebase. Let's think about how this could look like.
1.0
Populate conformance suite with "interesting" test cases - Our current methodology for testing the reference implementation involves writing a minimal amount of tests (https://github.com/openxla/stablehlo/blob/main/docs/reference_checklist.md), but for a conformance suite we'll likely want to write "interesting" tests that cover potentially surprising behavior, even if strictly speaking these tests don't increase coverage of the codebase. Let's think about how this could look like.
process
populate conformance suite with interesting test cases our current methodology for testing the reference implementation involves writing a minimal amount of tests but for a conformance suite we ll likely want to write interesting tests that cover potentially surprising behavior even if strictly speaking these tests don t increase coverage of the codebase let s think about how this could look like
1