Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
14,771
| 18,048,665,756
|
IssuesEvent
|
2021-09-19 10:54:00
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
MongoDB cannot use filter expression comparing two fields
|
Type:Bug Priority:P2 Database/Mongo Querying/Processor .Correctness
|
**Describe the bug**
When using Custom Expression filter on Mongo to compare two fields, then it always returns "No results".
Example `[field1] > [field2]` or `startsWith([field3], [field4])`
It correctly work if the second field is a static value like `[field1] > 123`.
Workaround is to use Native query and create `$match` which uses `$expr` references both fields.
**To Reproduce**
1. Custom question > **Mongo** Sample Dataset > Orders
2. Filter > Custom Expression > `[Discount] > [Quantity]`
3. Summarize Count

The query generated:
```json
[
{
"$match": {
"discount": {
"$gt": "$quantity"
}
}
},
{
"$group": {
"_id": null,
"count": {
"$sum": 1
}
}
},
{
"$sort": {
"_id": 1
}
},
{
"$project": {
"_id": false,
"count": true
}
}
]
```
The correct query should be using `$expr` - https://docs.mongodb.com/v4.0/reference/operator/query/expr/:
```json
[
{
"$match": {
"$expr": {
"$gt": [
"$discount",
"$quantity"
]
}
}
},
{
"$group": {
"_id": null,
"count": {
"$sum": 1
}
}
},
{
"$sort": {
"_id": 1
}
},
{
"$project": {
"_id": false,
"count": true
}
}
]
```
**Expected behavior**
Expecting 1337 results, similar to any other Sample Dataset

**Information about your Metabase Installation:**
Tested 0.37.8 thru 0.40.4 and master `43cf431`
|
1.0
|
MongoDB cannot use filter expression comparing two fields - **Describe the bug**
When using Custom Expression filter on Mongo to compare two fields, then it always returns "No results".
Example `[field1] > [field2]` or `startsWith([field3], [field4])`
It correctly work if the second field is a static value like `[field1] > 123`.
Workaround is to use Native query and create `$match` which uses `$expr` references both fields.
**To Reproduce**
1. Custom question > **Mongo** Sample Dataset > Orders
2. Filter > Custom Expression > `[Discount] > [Quantity]`
3. Summarize Count

The query generated:
```json
[
{
"$match": {
"discount": {
"$gt": "$quantity"
}
}
},
{
"$group": {
"_id": null,
"count": {
"$sum": 1
}
}
},
{
"$sort": {
"_id": 1
}
},
{
"$project": {
"_id": false,
"count": true
}
}
]
```
The correct query should be using `$expr` - https://docs.mongodb.com/v4.0/reference/operator/query/expr/:
```json
[
{
"$match": {
"$expr": {
"$gt": [
"$discount",
"$quantity"
]
}
}
},
{
"$group": {
"_id": null,
"count": {
"$sum": 1
}
}
},
{
"$sort": {
"_id": 1
}
},
{
"$project": {
"_id": false,
"count": true
}
}
]
```
**Expected behavior**
Expecting 1337 results, similar to any other Sample Dataset

**Information about your Metabase Installation:**
Tested 0.37.8 thru 0.40.4 and master `43cf431`
|
process
|
mongodb cannot use filter expression comparing two fields describe the bug when using custom expression filter on mongo to compare two fields then it always returns no results example or startswith it correctly work if the second field is a static value like workaround is to use native query and create match which uses expr references both fields to reproduce custom question mongo sample dataset orders filter custom expression summarize count the query generated json match discount gt quantity group id null count sum sort id project id false count true the correct query should be using expr json match expr gt discount quantity group id null count sum sort id project id false count true expected behavior expecting results similar to any other sample dataset information about your metabase installation tested thru and master
| 1
|
38,522
| 15,715,488,177
|
IssuesEvent
|
2021-03-28 01:34:45
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
7.12 Upgrade migrations fail if the advanced setting "timepicker:quickRanges" is null
|
Team:AppServices bug
|
https://github.com/elastic/kibana/pull/93409 Introduced a 7.12 migration for the `config` saved object that stores advanced settings. But if the advanced settings contains `"timepicker:quickRanges": null` the migration fails with logs like:
> Error: Failed to transform document 7.7.0. Transform: config:7.12.0
Doc: {"type":"config","id":"7.7.0","attributes":{"buildNum":30810,"defaultIndex":"8864-b38e6d72e23e...","dateFormat:dow":"Monday","state:storeInSessionStorage":true,"timepicker:quickRanges":null,"search:includeFrozen":null,"theme:darkMode":true,"courier:ignoreFilterIfFieldNotInIndex":true,"courier:batchSearches":null},"references":[],"migrationVersion":{"config":"7.9.0"},"updated_at":"2020-05-15T15:40:07.716Z"}
at tryTransformDoc (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:562:13)
at migrateProp (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:631:22)
at applyMigrations (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:350:20)
>
TypeError: Cannot read property 'indexOf' of null at /usr/share/kibana/src/core/server/ui_settings/saved_objects/migrations.js:29:69 at Array.reduce (<anonymous>) at 7.12.0 (/usr/share/kibana/src/core/server/ui_settings/saved_objects/migrations.js:28:47) at tryTransformDoc (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:547:22) at migrateProp (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:631:22) at applyMigrations (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:350:20)
|
1.0
|
7.12 Upgrade migrations fail if the advanced setting "timepicker:quickRanges" is null - https://github.com/elastic/kibana/pull/93409 Introduced a 7.12 migration for the `config` saved object that stores advanced settings. But if the advanced settings contains `"timepicker:quickRanges": null` the migration fails with logs like:
> Error: Failed to transform document 7.7.0. Transform: config:7.12.0
Doc: {"type":"config","id":"7.7.0","attributes":{"buildNum":30810,"defaultIndex":"8864-b38e6d72e23e...","dateFormat:dow":"Monday","state:storeInSessionStorage":true,"timepicker:quickRanges":null,"search:includeFrozen":null,"theme:darkMode":true,"courier:ignoreFilterIfFieldNotInIndex":true,"courier:batchSearches":null},"references":[],"migrationVersion":{"config":"7.9.0"},"updated_at":"2020-05-15T15:40:07.716Z"}
at tryTransformDoc (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:562:13)
at migrateProp (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:631:22)
at applyMigrations (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:350:20)
>
TypeError: Cannot read property 'indexOf' of null at /usr/share/kibana/src/core/server/ui_settings/saved_objects/migrations.js:29:69 at Array.reduce (<anonymous>) at 7.12.0 (/usr/share/kibana/src/core/server/ui_settings/saved_objects/migrations.js:28:47) at tryTransformDoc (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:547:22) at migrateProp (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:631:22) at applyMigrations (/usr/share/kibana/src/core/server/saved_objects/migrations/core/document_migrator.js:350:20)
|
non_process
|
upgrade migrations fail if the advanced setting timepicker quickranges is null introduced a migration for the config saved object that stores advanced settings but if the advanced settings contains timepicker quickranges null the migration fails with logs like error failed to transform document transform config doc type config id attributes buildnum defaultindex dateformat dow monday state storeinsessionstorage true timepicker quickranges null search includefrozen null theme darkmode true courier ignorefilteriffieldnotinindex true courier batchsearches null references migrationversion config updated at at trytransformdoc usr share kibana src core server saved objects migrations core document migrator js at migrateprop usr share kibana src core server saved objects migrations core document migrator js at applymigrations usr share kibana src core server saved objects migrations core document migrator js typeerror cannot read property indexof of null at usr share kibana src core server ui settings saved objects migrations js at array reduce at usr share kibana src core server ui settings saved objects migrations js at trytransformdoc usr share kibana src core server saved objects migrations core document migrator js at migrateprop usr share kibana src core server saved objects migrations core document migrator js at applymigrations usr share kibana src core server saved objects migrations core document migrator js
| 0
|
26,159
| 4,593,647,577
|
IssuesEvent
|
2016-09-21 02:13:44
|
afisher1/GridLAB-D
|
https://api.github.com/repos/afisher1/GridLAB-D
|
closed
|
#98 Meter not accruing power on energy,
|
defect
|
The powerflow meter is not properly accumulating the power into the energy property. Power is being recorded, but is simply not being accumulated into energy.
,
|
1.0
|
#98 Meter not accruing power on energy,
- The powerflow meter is not properly accumulating the power into the energy property. Power is being recorded, but is simply not being accumulated into energy.
,
|
non_process
|
meter not accruing power on energy the powerflow meter is not properly accumulating the power into the energy property power is being recorded but is simply not being accumulated into energy
| 0
|
184,454
| 14,289,346,824
|
IssuesEvent
|
2020-11-23 19:06:43
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
orijtech/frontender: lively/lively_test.go; 7 LoC
|
fresh test tiny
|
Found a possible issue in [orijtech/frontender](https://www.github.com/orijtech/frontender) at [lively/lively_test.go](https://github.com/orijtech/frontender/blob/34b58f87c92fefa14ed717149a0502b29c090d1f/lively/lively_test.go#L55-L61)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/orijtech/frontender/blob/34b58f87c92fefa14ed717149a0502b29c090d1f/lively/lively_test.go#L55-L61)
<details>
<summary>Click here to show the 7 line(s) of Go which triggered the analyzer.</summary>
```go
for i, peer := range peers {
waitCount += 1
go func(id int) {
_, _, err := peer.Liveliness(nil)
doneChan <- &peerDesc{i: id, err: err}
}(i)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 34b58f87c92fefa14ed717149a0502b29c090d1f
|
1.0
|
orijtech/frontender: lively/lively_test.go; 7 LoC -
Found a possible issue in [orijtech/frontender](https://www.github.com/orijtech/frontender) at [lively/lively_test.go](https://github.com/orijtech/frontender/blob/34b58f87c92fefa14ed717149a0502b29c090d1f/lively/lively_test.go#L55-L61)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/orijtech/frontender/blob/34b58f87c92fefa14ed717149a0502b29c090d1f/lively/lively_test.go#L55-L61)
<details>
<summary>Click here to show the 7 line(s) of Go which triggered the analyzer.</summary>
```go
for i, peer := range peers {
waitCount += 1
go func(id int) {
_, _, err := peer.Liveliness(nil)
doneChan <- &peerDesc{i: id, err: err}
}(i)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 34b58f87c92fefa14ed717149a0502b29c090d1f
|
non_process
|
orijtech frontender lively lively test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for i peer range peers waitcount go func id int err peer liveliness nil donechan peerdesc i id err err i leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
551,573
| 16,176,637,403
|
IssuesEvent
|
2021-05-03 07:59:59
|
NikolaiVChr/f16
|
https://api.github.com/repos/NikolaiVChr/f16
|
opened
|
Cockpit: Restore UHF panel to frequency selector instead of NAV.
|
Enhancement Low priority
|
Since the NAV freq can be tuned via DED and Mumble might soon b able to work through FG frequencies it would be good to restore the comm functionality.
|
1.0
|
Cockpit: Restore UHF panel to frequency selector instead of NAV. - Since the NAV freq can be tuned via DED and Mumble might soon b able to work through FG frequencies it would be good to restore the comm functionality.
|
non_process
|
cockpit restore uhf panel to frequency selector instead of nav since the nav freq can be tuned via ded and mumble might soon b able to work through fg frequencies it would be good to restore the comm functionality
| 0
|
224,155
| 17,153,369,381
|
IssuesEvent
|
2021-07-14 01:19:57
|
11ty/eleventy
|
https://api.github.com/repos/11ty/eleventy
|
closed
|
Serverless .gitignore example causes looping in --serve
|
documentation feature: π serverless
|
When following the example and [adding a .gitignore](https://www.11ty.dev/docs/plugins/serverless/#step-2-add-to-.gitignore), the `eleventy --serve` command will continue watching files that exist inside the `netlify/functions/possum/*` directory and cause a loop of file creation/watching/creation/etc.
If it's changed to `netlify/functions/possum/` the looping stops.
Not sure if there's a deeper underlying issue or if the docs just need to be updated.
|
1.0
|
Serverless .gitignore example causes looping in --serve - When following the example and [adding a .gitignore](https://www.11ty.dev/docs/plugins/serverless/#step-2-add-to-.gitignore), the `eleventy --serve` command will continue watching files that exist inside the `netlify/functions/possum/*` directory and cause a loop of file creation/watching/creation/etc.
If it's changed to `netlify/functions/possum/` the looping stops.
Not sure if there's a deeper underlying issue or if the docs just need to be updated.
|
non_process
|
serverless gitignore example causes looping in serve when following the example and the eleventy serve command will continue watching files that exist inside the netlify functions possum directory and cause a loop of file creation watching creation etc if it s changed to netlify functions possum the looping stops not sure if there s a deeper underlying issue or if the docs just need to be updated
| 0
|
79,722
| 7,723,939,749
|
IssuesEvent
|
2018-05-24 13:51:33
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: upreplicate/1to3 failed on master
|
C-test-failure O-robot
|
SHA: https://github.com/cockroachdb/cockroach/commits/9c197a359cf35d495425365cd22c3169e8cc0335
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=676291&tab=buildLog
```
cluster.go:468: /home/agent/work/.go/bin/roachprod create teamcity-676291-upreplicate-1to3 -n 3 --gce-machine-type=n1-standard-4: exit status 1
```
|
1.0
|
roachtest: upreplicate/1to3 failed on master - SHA: https://github.com/cockroachdb/cockroach/commits/9c197a359cf35d495425365cd22c3169e8cc0335
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=676291&tab=buildLog
```
cluster.go:468: /home/agent/work/.go/bin/roachprod create teamcity-676291-upreplicate-1to3 -n 3 --gce-machine-type=n1-standard-4: exit status 1
```
|
non_process
|
roachtest upreplicate failed on master sha parameters failed test cluster go home agent work go bin roachprod create teamcity upreplicate n gce machine type standard exit status
| 0
|
146,895
| 13,195,860,823
|
IssuesEvent
|
2020-08-13 19:29:36
|
marq24/UUID0xFD6FTracer
|
https://api.github.com/repos/marq24/UUID0xFD6FTracer
|
closed
|
Einstellungen
|
documentation question
|
Hallo,
welche Covid-Apps und welche LΓ€nder ausser Frankreich umfassen die Suche in den Einstellungen 1 und 3 der neuen Version 0.9.1.5 ?
Danke fΓΌr die groΓartige App
GrΓΌΓe
|
1.0
|
Einstellungen - Hallo,
welche Covid-Apps und welche LΓ€nder ausser Frankreich umfassen die Suche in den Einstellungen 1 und 3 der neuen Version 0.9.1.5 ?
Danke fΓΌr die groΓartige App
GrΓΌΓe
|
non_process
|
einstellungen hallo welche covid apps und welche lΓ€nder ausser frankreich umfassen die suche in den einstellungen und der neuen version danke fΓΌr die groΓartige app grΓΌΓe
| 0
|
2,849
| 5,809,459,246
|
IssuesEvent
|
2017-05-04 13:28:24
|
P0cL4bs/WiFi-Pumpkin
|
https://api.github.com/repos/P0cL4bs/WiFi-Pumpkin
|
closed
|
Python DNS Server improvements
|
enhancement help wanted in process priority solved
|
I need help for do test the new pyDNS Server implementation. the current version 0.8.4 is running the dns2proxy because i had some problems with DNS server and i was forced to solve this problem temporarily last commit. so i doing new test with new version the DNS server for add new features.
|
1.0
|
Python DNS Server improvements - I need help for do test the new pyDNS Server implementation. the current version 0.8.4 is running the dns2proxy because i had some problems with DNS server and i was forced to solve this problem temporarily last commit. so i doing new test with new version the DNS server for add new features.
|
process
|
python dns server improvements i need help for do test the new pydns server implementation the current version is running the because i had some problems with dns server and i was forced to solve this problem temporarily last commit so i doing new test with new version the dns server for add new features
| 1
|
85,574
| 16,675,059,193
|
IssuesEvent
|
2021-06-07 15:14:39
|
apl-cornell/PDL
|
https://api.github.com/repos/apl-cornell/PDL
|
closed
|
Implement Speculative State
|
code generation enhancement
|
In order to properly track speculation, each pipeline stage needs to keep track of an identifier which represents the current instruction's reference into a "Speculation Table".
This issue tracks implementing:
- Commands for checking a stage's current speculative state (blocking & non-blocking)
- Code Generation for Killing a stage's current execution (i.e., a rule that reads inputs, but does nothing with them)
This will need to be integrated with #15 that tracks implementing speculative calls.
|
1.0
|
Implement Speculative State - In order to properly track speculation, each pipeline stage needs to keep track of an identifier which represents the current instruction's reference into a "Speculation Table".
This issue tracks implementing:
- Commands for checking a stage's current speculative state (blocking & non-blocking)
- Code Generation for Killing a stage's current execution (i.e., a rule that reads inputs, but does nothing with them)
This will need to be integrated with #15 that tracks implementing speculative calls.
|
non_process
|
implement speculative state in order to properly track speculation each pipeline stage needs to keep track of an identifier which represents the current instruction s reference into a speculation table this issue tracks implementing commands for checking a stage s current speculative state blocking non blocking code generation for killing a stage s current execution i e a rule that reads inputs but does nothing with them this will need to be integrated with that tracks implementing speculative calls
| 0
|
502,989
| 14,576,904,256
|
IssuesEvent
|
2020-12-18 00:37:16
|
LinkedEarth/Pyleoclim_util
|
https://api.github.com/repos/LinkedEarth/Pyleoclim_util
|
closed
|
Align parameters for MutipleSeries.spectral and Series.Spectral
|
low priority
|
**Is your feature request related to a problem? Please describe.**
Parameters between the two functions are different (confusing).
**Describe the solution you'd like**
Agree on the set of parameters that should be exposed and align the two
**Describe alternatives you've considered**
leave as is
**Additional context**
NA
|
1.0
|
Align parameters for MutipleSeries.spectral and Series.Spectral - **Is your feature request related to a problem? Please describe.**
Parameters between the two functions are different (confusing).
**Describe the solution you'd like**
Agree on the set of parameters that should be exposed and align the two
**Describe alternatives you've considered**
leave as is
**Additional context**
NA
|
non_process
|
align parameters for mutipleseries spectral and series spectral is your feature request related to a problem please describe parameters between the two functions are different confusing describe the solution you d like agree on the set of parameters that should be exposed and align the two describe alternatives you ve considered leave as is additional context na
| 0
|
14,036
| 16,843,896,960
|
IssuesEvent
|
2021-06-19 04:01:35
|
googleapis/repo-automation-bots
|
https://api.github.com/repos/googleapis/repo-automation-bots
|
closed
|
confirm GitHub signature is being validated in gcf-utils
|
type: process
|
@orthros I noticed debugging our tasks work, that we were populating the wrong header for `x-github-signature`, and logic was still working:
see: https://github.com/googleapis/repo-automation-bots/pull/469
We should add tests that confirm the signature validation step gates webhooks.
|
1.0
|
confirm GitHub signature is being validated in gcf-utils - @orthros I noticed debugging our tasks work, that we were populating the wrong header for `x-github-signature`, and logic was still working:
see: https://github.com/googleapis/repo-automation-bots/pull/469
We should add tests that confirm the signature validation step gates webhooks.
|
process
|
confirm github signature is being validated in gcf utils orthros i noticed debugging our tasks work that we were populating the wrong header for x github signature and logic was still working see we should add tests that confirm the signature validation step gates webhooks
| 1
|
16,968
| 22,331,410,060
|
IssuesEvent
|
2022-06-14 14:47:40
|
opensafely-core/job-server
|
https://api.github.com/repos/opensafely-core/job-server
|
opened
|
Type of phone label not working
|
application-process
|
type of phone not displaying content correctly


|
1.0
|
Type of phone label not working - type of phone not displaying content correctly


|
process
|
type of phone label not working type of phone not displaying content correctly
| 1
|
44,867
| 23,798,193,593
|
IssuesEvent
|
2022-09-02 23:24:23
|
rapidsai/cudf
|
https://api.github.com/repos/rapidsai/cudf
|
opened
|
Optimize `to_cupy` and `values`
|
cuDF (Python) Performance improvement
|
Currently `series.values` and especially `series.to_cupy()` are substantially slower than `cupy.asarray(series)`.
```
In [2]: s = cudf.Series(range(10000))
In [3]: %timeit s.values
81.4 Β΅s Β± 1.68 Β΅s per loop (mean Β± std. dev. of 7 runs, 10,000 loops each)
In [4]: %timeit cp.asarray(s)
19.1 Β΅s Β± 168 ns per loop (mean Β± std. dev. of 7 runs, 100,000 loops each)
In [5]: %timeit s.to_cupy()
349 Β΅s Β± 75.2 Β΅s per loop (mean Β± std. dev. of 7 runs, 1 loop each)
```
There are at least two obvious potential culprits in `Frame._to_array` (the underlying method for `to_cupy`):
- [It always performs an extra allocation](https://github.com/rapidsai/cudf/blob/branch-22.10/python/cudf/cudf/core/frame.py#L484), even when `copy=False`.
- [It performs dtype inference using `find_common_dtype`](https://github.com/rapidsai/cudf/blob/branch-22.10/python/cudf/cudf/core/frame.py#L479), which is _slow_ (and slower for `DataFrame`s with many columns):
```
In [11]: df = cudf.DataFrame({'a': [1], 'b': [3.], 'c': ['a']})
In [12]: %timeit cudf.utils.dtypes.find_common_type([col.dtype for col in df._data.values()])
53.6 Β΅s Β± 530 ns per loop (mean Β± std. dev. of 7 runs, 10,000 loops each)
In [13]: df = cudf.DataFrame({'a': [1], 'b': [3.]})
In [14]: %timeit cudf.utils.dtypes.find_common_type([col.dtype for col in df._data.values()])
39.8 Β΅s Β± 1.01 Β΅s per loop (mean Β± std. dev. of 7 runs, 10,000 loops each)
```
The implementation of `values` drops down to `ColumnBase.values` and requires some deeper consideration. However, since we use `.values` frequently internally (and we occasionally use `to_cupy`) we are likely giving up a lot of performance. We should profile these functions to determine the bottlenecks, and if there are valid reasons for them we should establish some policies on how to select the right function to use when performing these conversions to arrays internally. While this exact analogy does not hold for `DataFrame` (because that doesn't support the conversion to an array), any optimization that we make for `Series` will likely also help speed up `DataFrame` operations.
|
True
|
Optimize `to_cupy` and `values` - Currently `series.values` and especially `series.to_cupy()` are substantially slower than `cupy.asarray(series)`.
```
In [2]: s = cudf.Series(range(10000))
In [3]: %timeit s.values
81.4 Β΅s Β± 1.68 Β΅s per loop (mean Β± std. dev. of 7 runs, 10,000 loops each)
In [4]: %timeit cp.asarray(s)
19.1 Β΅s Β± 168 ns per loop (mean Β± std. dev. of 7 runs, 100,000 loops each)
In [5]: %timeit s.to_cupy()
349 Β΅s Β± 75.2 Β΅s per loop (mean Β± std. dev. of 7 runs, 1 loop each)
```
There are at least two obvious potential culprits in `Frame._to_array` (the underlying method for `to_cupy`):
- [It always performs an extra allocation](https://github.com/rapidsai/cudf/blob/branch-22.10/python/cudf/cudf/core/frame.py#L484), even when `copy=False`.
- [It performs dtype inference using `find_common_dtype`](https://github.com/rapidsai/cudf/blob/branch-22.10/python/cudf/cudf/core/frame.py#L479), which is _slow_ (and slower for `DataFrame`s with many columns):
```
In [11]: df = cudf.DataFrame({'a': [1], 'b': [3.], 'c': ['a']})
In [12]: %timeit cudf.utils.dtypes.find_common_type([col.dtype for col in df._data.values()])
53.6 Β΅s Β± 530 ns per loop (mean Β± std. dev. of 7 runs, 10,000 loops each)
In [13]: df = cudf.DataFrame({'a': [1], 'b': [3.]})
In [14]: %timeit cudf.utils.dtypes.find_common_type([col.dtype for col in df._data.values()])
39.8 Β΅s Β± 1.01 Β΅s per loop (mean Β± std. dev. of 7 runs, 10,000 loops each)
```
The implementation of `values` drops down to `ColumnBase.values` and requires some deeper consideration. However, since we use `.values` frequently internally (and we occasionally use `to_cupy`) we are likely giving up a lot of performance. We should profile these functions to determine the bottlenecks, and if there are valid reasons for them we should establish some policies on how to select the right function to use when performing these conversions to arrays internally. While this exact analogy does not hold for `DataFrame` (because that doesn't support the conversion to an array), any optimization that we make for `Series` will likely also help speed up `DataFrame` operations.
|
non_process
|
optimize to cupy and values currently series values and especially series to cupy are substantially slower than cupy asarray series in s cudf series range in timeit s values Β΅s Β± Β΅s per loop mean Β± std dev of runs loops each in timeit cp asarray s Β΅s Β± ns per loop mean Β± std dev of runs loops each in timeit s to cupy Β΅s Β± Β΅s per loop mean Β± std dev of runs loop each there are at least two obvious potential culprits in frame to array the underlying method for to cupy even when copy false which is slow and slower for dataframe s with many columns in df cudf dataframe a b c in timeit cudf utils dtypes find common type Β΅s Β± ns per loop mean Β± std dev of runs loops each in df cudf dataframe a b in timeit cudf utils dtypes find common type Β΅s Β± Β΅s per loop mean Β± std dev of runs loops each the implementation of values drops down to columnbase values and requires some deeper consideration however since we use values frequently internally and we occasionally use to cupy we are likely giving up a lot of performance we should profile these functions to determine the bottlenecks and if there are valid reasons for them we should establish some policies on how to select the right function to use when performing these conversions to arrays internally while this exact analogy does not hold for dataframe because that doesn t support the conversion to an array any optimization that we make for series will likely also help speed up dataframe operations
| 0
|
237,919
| 7,768,274,970
|
IssuesEvent
|
2018-06-03 16:14:42
|
jahirfiquitiva/Blueprint
|
https://api.github.com/repos/jahirfiquitiva/Blueprint
|
closed
|
Blueprint crashes while opening wallpapers from wallpapers tab on android 5.0
|
Priority: High Status: Accepted Type: Bug
|
<!--
Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them.
Put an x inside the [] like this: [x] to mark the checkbox.
-->
- [x] I have verified there are no duplicate active or recent bugs, questions, or requests
- [x] I have verified that I am using the latest version of Blueprint.
### Device/dashboard info:
- Blueprint version: `1.1.7 & 1.1.5`
- Android version: `5.0`
- Device Manufacturer: `Various`
- Device Name: `AVD Emulator AOSP, P6, Idol 3, ZenFone 2, Galaxy Alpha, Galaxy Note 3, Canvas Unite 2, K920/Vibe, Galaxy S4, Glade V6, Cloud Power+, x98air3g, Canvas Fire 4, C1AE-2, Redmi Note, ZenFone 2 Laser, Galaxy Tab A, LG Leon, Galaxy Note 3`
```Gradle
java.lang.NullPointerException:
at android.view.ViewOverlay$OverlayViewGroup.add (ViewOverlay.java:164)
at android.view.ViewGroupOverlay.add (ViewGroupOverlay.java:63)
at android.app.EnterTransitionCoordinator.startRejectedAnimations (EnterTransitionCoordinator.java:598)
at android.app.EnterTransitionCoordinator.startSharedElementTransition (EnterTransitionCoordinator.java:325)
at android.app.EnterTransitionCoordinator.access$200 (EnterTransitionCoordinator.java:42)
at android.app.EnterTransitionCoordinator$5$1.run (EnterTransitionCoordinator.java:389)
at android.app.ActivityTransitionCoordinator.startTransition (ActivityTransitionCoordinator.java:698)
at android.app.EnterTransitionCoordinator$5.onPreDraw (EnterTransitionCoordinator.java:386)
at android.view.ViewTreeObserver.dispatchOnPreDraw (ViewTreeObserver.java:847)
at android.view.ViewRootImpl.performTraversals (ViewRootImpl.java:2294)
at android.view.ViewRootImpl.doTraversal (ViewRootImpl.java:1251)
at android.view.ViewRootImpl$TraversalRunnable.run (ViewRootImpl.java:6438)
at android.view.Choreographer$CallbackRecord.run (Choreographer.java:795)
at android.view.Choreographer.doCallbacks (Choreographer.java:598)
at android.view.Choreographer.doFrame (Choreographer.java:567)
at android.view.Choreographer$FrameDisplayEventReceiver.run (Choreographer.java:781)
at android.os.Handler.handleCallback (Handler.java:810)
at android.os.Handler.dispatchMessage (Handler.java:99)
at android.os.Looper.loop (Looper.java:189)
at android.app.ActivityThread.main (ActivityThread.java:5529)
at java.lang.reflect.Method.invoke (Native Method)
at java.lang.reflect.Method.invoke (Method.java:372)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run (ZygoteInit.java:950)
at com.android.internal.os.ZygoteInit.main (ZygoteInit.java:745)
```
<!--
The required steps to reproduce it.
-->
### Reproduction Steps
1. Open any wallpaper from wallpapers tab
### Expected Result
Wallpaper view opens
### Actual Result
App crashes
|
1.0
|
Blueprint crashes while opening wallpapers from wallpapers tab on android 5.0 - <!--
Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them.
Put an x inside the [] like this: [x] to mark the checkbox.
-->
- [x] I have verified there are no duplicate active or recent bugs, questions, or requests
- [x] I have verified that I am using the latest version of Blueprint.
### Device/dashboard info:
- Blueprint version: `1.1.7 & 1.1.5`
- Android version: `5.0`
- Device Manufacturer: `Various`
- Device Name: `AVD Emulator AOSP, P6, Idol 3, ZenFone 2, Galaxy Alpha, Galaxy Note 3, Canvas Unite 2, K920/Vibe, Galaxy S4, Glade V6, Cloud Power+, x98air3g, Canvas Fire 4, C1AE-2, Redmi Note, ZenFone 2 Laser, Galaxy Tab A, LG Leon, Galaxy Note 3`
```Gradle
java.lang.NullPointerException:
at android.view.ViewOverlay$OverlayViewGroup.add (ViewOverlay.java:164)
at android.view.ViewGroupOverlay.add (ViewGroupOverlay.java:63)
at android.app.EnterTransitionCoordinator.startRejectedAnimations (EnterTransitionCoordinator.java:598)
at android.app.EnterTransitionCoordinator.startSharedElementTransition (EnterTransitionCoordinator.java:325)
at android.app.EnterTransitionCoordinator.access$200 (EnterTransitionCoordinator.java:42)
at android.app.EnterTransitionCoordinator$5$1.run (EnterTransitionCoordinator.java:389)
at android.app.ActivityTransitionCoordinator.startTransition (ActivityTransitionCoordinator.java:698)
at android.app.EnterTransitionCoordinator$5.onPreDraw (EnterTransitionCoordinator.java:386)
at android.view.ViewTreeObserver.dispatchOnPreDraw (ViewTreeObserver.java:847)
at android.view.ViewRootImpl.performTraversals (ViewRootImpl.java:2294)
at android.view.ViewRootImpl.doTraversal (ViewRootImpl.java:1251)
at android.view.ViewRootImpl$TraversalRunnable.run (ViewRootImpl.java:6438)
at android.view.Choreographer$CallbackRecord.run (Choreographer.java:795)
at android.view.Choreographer.doCallbacks (Choreographer.java:598)
at android.view.Choreographer.doFrame (Choreographer.java:567)
at android.view.Choreographer$FrameDisplayEventReceiver.run (Choreographer.java:781)
at android.os.Handler.handleCallback (Handler.java:810)
at android.os.Handler.dispatchMessage (Handler.java:99)
at android.os.Looper.loop (Looper.java:189)
at android.app.ActivityThread.main (ActivityThread.java:5529)
at java.lang.reflect.Method.invoke (Native Method)
at java.lang.reflect.Method.invoke (Method.java:372)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run (ZygoteInit.java:950)
at com.android.internal.os.ZygoteInit.main (ZygoteInit.java:745)
```
<!--
The required steps to reproduce it.
-->
### Reproduction Steps
1. Open any wallpaper from wallpapers tab
### Expected Result
Wallpaper view opens
### Actual Result
App crashes
|
non_process
|
blueprint crashes while opening wallpapers from wallpapers tab on android any html comment will be stripped when the markdown is rendered so you don t need to delete them put an x inside the like this to mark the checkbox i have verified there are no duplicate active or recent bugs questions or requests i have verified that i am using the latest version of blueprint device dashboard info blueprint version android version device manufacturer various device name avd emulator aosp idol zenfone galaxy alpha galaxy note canvas unite vibe galaxy glade cloud power canvas fire redmi note zenfone laser galaxy tab a lg leon galaxy note gradle java lang nullpointerexception at android view viewoverlay overlayviewgroup add viewoverlay java at android view viewgroupoverlay add viewgroupoverlay java at android app entertransitioncoordinator startrejectedanimations entertransitioncoordinator java at android app entertransitioncoordinator startsharedelementtransition entertransitioncoordinator java at android app entertransitioncoordinator access entertransitioncoordinator java at android app entertransitioncoordinator run entertransitioncoordinator java at android app activitytransitioncoordinator starttransition activitytransitioncoordinator java at android app entertransitioncoordinator onpredraw entertransitioncoordinator java at android view viewtreeobserver dispatchonpredraw viewtreeobserver java at android view viewrootimpl performtraversals viewrootimpl java at android view viewrootimpl dotraversal viewrootimpl java at android view viewrootimpl traversalrunnable run viewrootimpl java at android view choreographer callbackrecord run choreographer java at android view choreographer docallbacks choreographer java at android view choreographer doframe choreographer java at android view choreographer framedisplayeventreceiver run choreographer java at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java the required steps to reproduce it reproduction steps open any wallpaper from wallpapers tab expected result wallpaper view opens actual result app crashes
| 0
|
633,577
| 20,259,166,196
|
IssuesEvent
|
2022-02-15 04:37:51
|
therealbluepandabear/PyxlMoose
|
https://api.github.com/repos/therealbluepandabear/PyxlMoose
|
closed
|
[Feature Request] Spray tool
|
β¨ enhancement low priority
|
#### Feature description
This feature will allow to use to emulate how a spray paint works in real life, but on a pixel art canvas inside the app.
The user can edit options such as 'radius' or 'strength' to their liking.
#### Why is this feature important to add?
This feature is important to add as I've seen a couple of similar apps that have this feature, and it can be used when generating random pieces of 'information' is needed.
|
1.0
|
[Feature Request] Spray tool - #### Feature description
This feature will allow to use to emulate how a spray paint works in real life, but on a pixel art canvas inside the app.
The user can edit options such as 'radius' or 'strength' to their liking.
#### Why is this feature important to add?
This feature is important to add as I've seen a couple of similar apps that have this feature, and it can be used when generating random pieces of 'information' is needed.
|
non_process
|
spray tool feature description this feature will allow to use to emulate how a spray paint works in real life but on a pixel art canvas inside the app the user can edit options such as radius or strength to their liking why is this feature important to add this feature is important to add as i ve seen a couple of similar apps that have this feature and it can be used when generating random pieces of information is needed
| 0
|
5,805
| 8,643,540,775
|
IssuesEvent
|
2018-11-25 18:55:12
|
gfrebello/qs-trip-planning-procedure
|
https://api.github.com/repos/gfrebello/qs-trip-planning-procedure
|
closed
|
Integrate front end for flight reservation with the database
|
Priority:High Process:Implement Requirement
|
Once the front end is done, and the entities for flights are created, and a few flights are added to the database, the integration between these parts need to be made. The front end needs to retrieve the flights from the database and display them to the user.
https://stackoverflow.com/questions/29042911/split-the-date-and-time-in-two-elements
https://developer.mozilla.org/pt-BR/docs/Web/JavaScript/Reference/Global_Objects/Array/sort
https://github.com/palantir/tslint/issues/1449
https://stackoverflow.com/questions/85992/how-do-i-enumerate-the-properties-of-a-javascript-object
https://stackoverflow.com/questions/40911194/how-do-i-add-an-element-to-array-in-reducer-of-react-native-redux
https://stackoverflow.com/questions/4566771/how-to-globally-replace-a-forward-slash-in-a-javascript-string
|
1.0
|
Integrate front end for flight reservation with the database - Once the front end is done, and the entities for flights are created, and a few flights are added to the database, the integration between these parts need to be made. The front end needs to retrieve the flights from the database and display them to the user.
https://stackoverflow.com/questions/29042911/split-the-date-and-time-in-two-elements
https://developer.mozilla.org/pt-BR/docs/Web/JavaScript/Reference/Global_Objects/Array/sort
https://github.com/palantir/tslint/issues/1449
https://stackoverflow.com/questions/85992/how-do-i-enumerate-the-properties-of-a-javascript-object
https://stackoverflow.com/questions/40911194/how-do-i-add-an-element-to-array-in-reducer-of-react-native-redux
https://stackoverflow.com/questions/4566771/how-to-globally-replace-a-forward-slash-in-a-javascript-string
|
process
|
integrate front end for flight reservation with the database once the front end is done and the entities for flights are created and a few flights are added to the database the integration between these parts need to be made the front end needs to retrieve the flights from the database and display them to the user
| 1
|
206,679
| 15,767,895,791
|
IssuesEvent
|
2021-03-31 16:36:04
|
influxdata/influxdb
|
https://api.github.com/repos/influxdata/influxdb
|
closed
|
Backport flux test harness to 1.x
|
1.x area/flux area/tests
|
The new flux test harness used by 2.x can assert that pushdowns are actually used. Backporting it to the `master-1.x` branch will help us be sure #21069 is complete.
|
1.0
|
Backport flux test harness to 1.x - The new flux test harness used by 2.x can assert that pushdowns are actually used. Backporting it to the `master-1.x` branch will help us be sure #21069 is complete.
|
non_process
|
backport flux test harness to x the new flux test harness used by x can assert that pushdowns are actually used backporting it to the master x branch will help us be sure is complete
| 0
|
237,883
| 26,085,429,233
|
IssuesEvent
|
2022-12-26 01:44:20
|
faizulho/gatsby-starter-docz-netlifycms-1
|
https://api.github.com/repos/faizulho/gatsby-starter-docz-netlifycms-1
|
opened
|
CVE-2022-37603 (High) detected in loader-utils-1.4.0.tgz
|
security vulnerability
|
## CVE-2022-37603 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-1.4.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-2.30.3.tgz (Root Library)
- babel-loader-8.2.2.tgz
- :x: **loader-utils-1.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/faizulho/gatsby-starter-docz-netlifycms-1/commit/70a9e87b1e68c0bef6964284e0899376209b0f3d">70a9e87b1e68c0bef6964284e0899376209b0f3d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js.
<p>Publish Date: 2022-10-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37603>CVE-2022-37603</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-3rfm-jhwj-7488">https://github.com/advisories/GHSA-3rfm-jhwj-7488</a></p>
<p>Release Date: 2022-10-14</p>
<p>Fix Resolution (loader-utils): 2.0.4</p>
<p>Direct dependency fix Resolution (gatsby): 2.31.0-incbuilds-special.22</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-37603 (High) detected in loader-utils-1.4.0.tgz - ## CVE-2022-37603 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-1.4.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-2.30.3.tgz (Root Library)
- babel-loader-8.2.2.tgz
- :x: **loader-utils-1.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/faizulho/gatsby-starter-docz-netlifycms-1/commit/70a9e87b1e68c0bef6964284e0899376209b0f3d">70a9e87b1e68c0bef6964284e0899376209b0f3d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js.
<p>Publish Date: 2022-10-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37603>CVE-2022-37603</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-3rfm-jhwj-7488">https://github.com/advisories/GHSA-3rfm-jhwj-7488</a></p>
<p>Release Date: 2022-10-14</p>
<p>Fix Resolution (loader-utils): 2.0.4</p>
<p>Direct dependency fix Resolution (gatsby): 2.31.0-incbuilds-special.22</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in loader utils tgz cve high severity vulnerability vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy gatsby tgz root library babel loader tgz x loader utils tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the url variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution loader utils direct dependency fix resolution gatsby incbuilds special step up your open source security game with mend
| 0
|
35,304
| 17,021,000,897
|
IssuesEvent
|
2021-07-02 19:02:50
|
fkk-cz/noire_vehicles
|
https://api.github.com/repos/fkk-cz/noire_vehicles
|
closed
|
mclaren 720s n-largo
|
performance issue
|
please make the acceleration better, the car really only starts going when its at 80mp/h for its price I definitely think it should have a better acceleration if not better acceleration and better top speed cuz it goes 180-190 mp/h on flat.
and be able to dobble clutch would be nice:D
|
True
|
mclaren 720s n-largo - please make the acceleration better, the car really only starts going when its at 80mp/h for its price I definitely think it should have a better acceleration if not better acceleration and better top speed cuz it goes 180-190 mp/h on flat.
and be able to dobble clutch would be nice:D
|
non_process
|
mclaren n largo please make the acceleration better the car really only starts going when its at h for its price i definitely think it should have a better acceleration if not better acceleration and better top speed cuz it goes mp h on flat and be able to dobble clutch would be nice d
| 0
|
7,900
| 11,089,087,525
|
IssuesEvent
|
2019-12-14 16:00:59
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Handling of the "DOTX012W" error message
|
preprocess/conref stale
|
If you look in the docs for this specific error message:
http://www.dita-ot.org/dev/user-guide/DITA-messages.html
it says that the message should not appear anymore. But it still does.
I'm attaching a sample project but I think you can obtain this easier if you a step conref from a standard DITA task to a DITA general task.
[testMessage.zip](https://github.com/dita-ot/dita-ot/files/1394534/testMessage.zip)
|
1.0
|
Handling of the "DOTX012W" error message - If you look in the docs for this specific error message:
http://www.dita-ot.org/dev/user-guide/DITA-messages.html
it says that the message should not appear anymore. But it still does.
I'm attaching a sample project but I think you can obtain this easier if you a step conref from a standard DITA task to a DITA general task.
[testMessage.zip](https://github.com/dita-ot/dita-ot/files/1394534/testMessage.zip)
|
process
|
handling of the error message if you look in the docs for this specific error message it says that the message should not appear anymore but it still does i m attaching a sample project but i think you can obtain this easier if you a step conref from a standard dita task to a dita general task
| 1
|
307,984
| 23,225,301,238
|
IssuesEvent
|
2022-08-02 23:01:16
|
tidyverse/dplyr
|
https://api.github.com/repos/tidyverse/dplyr
|
closed
|
Clearer across() documentation for multiple .fns
|
documentation each-col βοΈ
|
I've found the documentation for across() about the multiple function case more oblique than perhaps it could be, but I'm not familiar enough with all of its applications to be sure I'll get it right if I directly submit a change via pull. This applies to both the Roxygen comment on R/across.R and to a lesser degree the colwise vignette, vignettes/colwise.Rmd
The current text in the Roxygen comment is
>Functions to apply to each of the selected columns. Possible values are:
> - NULL, to returns the columns untransformed.
> - A function, e.g. mean.
> - A purrr-style lambda, e.g. ~ mean(.x, na.rm = TRUE)
> - A list of functions/lambdas, e.g. list(mean = mean, n_miss = ~ sum(is.na(.x))
and in the vignette,
> The second argument, .fns, is a function or list of functions to apply to each column. This can also be a purrr style formula (or list of formulas) like ~ .x / 2. (This argument is optional, and you can omit it if you just want to get the underlying data; youβll see that technique used in vignette("rowwise").)
> ...
> ### Multiple functions
>
> You can transform each variable with more than one function by supplying a named list of functions or lambda functions in the second argument:
The intuition that I'm only seeing referred to indirectly is that the provided functions *fan out* into multiple outputs, i.e. if you provided n columns and m functions you'd end up with n x m outputs. The Roxygen intro also refers to transformation in the singular, "across() makes it easy to apply the same transformation to multiple columns...", so it feels like the multiple-output case is de-emphasised. There are other possible interpretations for "Functions to apply to each of the selected columns", e.g. maybe the functions are applied in order, to a single output. (This is of course very achievable through other methods like `purrr::compose` or `. %>% ...` pipes or purrr lambdas, and not necessary for `across` to do, but it's not ruled out by the description.)
It's also not made entirely clear that an _unnamed_ list of functions `c(...)` is probably unusual, i.e. the reason a _named_ list is used is because the name determines the output name (depending on `.names`) - someone might use an unnamed list in an ad-hoc to test a range of transformations, but named lists match the intention behind the design of the function better. The vignette is clearer about this, and while Roxygen uses a named list as an example but leaves the choice unaddressed, i.e. it could just be a code style cue to make a lambda more readable. (I've primarily used across with `mutate`, so the rationale in the `summarise`-style example wasn't immediately obvious.)
My suggestion would be to emphasise that a separate output is produced for each function, e.g. a maybe clumsy alternative wording would be
> Functions to apply to each of the selected columns, where each transformation produces a separate output column based on the column and list names.
Again, I don't know for sure if this conflicts with any current or planned uses for the function. As far as I can tell across more or less enforces this as a constraint, i.e. requires unique names for each output (although of course transform functions can return NULL in order to _not_ produce a new output column. _Optionally_ producing output for a list of functions opens up some interesting variations I guess that vary a little from the n x m account.)
Mostly my sense was that the intention or purpose behind the argument was underspecified in the documentation, making it harder to understand the behaviour unless you already know the behaviour. Partly this could be a consequence of across being a fairly flexible tool so its 'meaning' can look a little different in different uses, but I think we can nudge the docs towards giving a quicker birds-eye view. Hopefully you can see where I'm coming from.
|
1.0
|
Clearer across() documentation for multiple .fns - I've found the documentation for across() about the multiple function case more oblique than perhaps it could be, but I'm not familiar enough with all of its applications to be sure I'll get it right if I directly submit a change via pull. This applies to both the Roxygen comment on R/across.R and to a lesser degree the colwise vignette, vignettes/colwise.Rmd
The current text in the Roxygen comment is
>Functions to apply to each of the selected columns. Possible values are:
> - NULL, to returns the columns untransformed.
> - A function, e.g. mean.
> - A purrr-style lambda, e.g. ~ mean(.x, na.rm = TRUE)
> - A list of functions/lambdas, e.g. list(mean = mean, n_miss = ~ sum(is.na(.x))
and in the vignette,
> The second argument, .fns, is a function or list of functions to apply to each column. This can also be a purrr style formula (or list of formulas) like ~ .x / 2. (This argument is optional, and you can omit it if you just want to get the underlying data; youβll see that technique used in vignette("rowwise").)
> ...
> ### Multiple functions
>
> You can transform each variable with more than one function by supplying a named list of functions or lambda functions in the second argument:
The intuition that I'm only seeing referred to indirectly is that the provided functions *fan out* into multiple outputs, i.e. if you provided n columns and m functions you'd end up with n x m outputs. The Roxygen intro also refers to transformation in the singular, "across() makes it easy to apply the same transformation to multiple columns...", so it feels like the multiple-output case is de-emphasised. There are other possible interpretations for "Functions to apply to each of the selected columns", e.g. maybe the functions are applied in order, to a single output. (This is of course very achievable through other methods like `purrr::compose` or `. %>% ...` pipes or purrr lambdas, and not necessary for `across` to do, but it's not ruled out by the description.)
It's also not made entirely clear that an _unnamed_ list of functions `c(...)` is probably unusual, i.e. the reason a _named_ list is used is because the name determines the output name (depending on `.names`) - someone might use an unnamed list in an ad-hoc to test a range of transformations, but named lists match the intention behind the design of the function better. The vignette is clearer about this, and while Roxygen uses a named list as an example but leaves the choice unaddressed, i.e. it could just be a code style cue to make a lambda more readable. (I've primarily used across with `mutate`, so the rationale in the `summarise`-style example wasn't immediately obvious.)
My suggestion would be to emphasise that a separate output is produced for each function, e.g. a maybe clumsy alternative wording would be
> Functions to apply to each of the selected columns, where each transformation produces a separate output column based on the column and list names.
Again, I don't know for sure if this conflicts with any current or planned uses for the function. As far as I can tell across more or less enforces this as a constraint, i.e. requires unique names for each output (although of course transform functions can return NULL in order to _not_ produce a new output column. _Optionally_ producing output for a list of functions opens up some interesting variations I guess that vary a little from the n x m account.)
Mostly my sense was that the intention or purpose behind the argument was underspecified in the documentation, making it harder to understand the behaviour unless you already know the behaviour. Partly this could be a consequence of across being a fairly flexible tool so its 'meaning' can look a little different in different uses, but I think we can nudge the docs towards giving a quicker birds-eye view. Hopefully you can see where I'm coming from.
|
non_process
|
clearer across documentation for multiple fns i ve found the documentation for across about the multiple function case more oblique than perhaps it could be but i m not familiar enough with all of its applications to be sure i ll get it right if i directly submit a change via pull this applies to both the roxygen comment on r across r and to a lesser degree the colwise vignette vignettes colwise rmd the current text in the roxygen comment is functions to apply to each of the selected columns possible values are null to returns the columns untransformed a function e g mean a purrr style lambda e g mean x na rm true a list of functions lambdas e g list mean mean n miss sum is na x and in the vignette the second argument fns is a function or list of functions to apply to each column this can also be a purrr style formula or list of formulas like x this argument is optional and you can omit it if you just want to get the underlying data youβll see that technique used in vignette rowwise multiple functions you can transform each variable with more than one function by supplying a named list of functions or lambda functions in the second argument the intuition that i m only seeing referred to indirectly is that the provided functions fan out into multiple outputs i e if you provided n columns and m functions you d end up with n x m outputs the roxygen intro also refers to transformation in the singular across makes it easy to apply the same transformation to multiple columns so it feels like the multiple output case is de emphasised there are other possible interpretations for functions to apply to each of the selected columns e g maybe the functions are applied in order to a single output this is of course very achievable through other methods like purrr compose or pipes or purrr lambdas and not necessary for across to do but it s not ruled out by the description it s also not made entirely clear that an unnamed list of functions c is probably unusual i e the reason a named list is used is because the name determines the output name depending on names someone might use an unnamed list in an ad hoc to test a range of transformations but named lists match the intention behind the design of the function better the vignette is clearer about this and while roxygen uses a named list as an example but leaves the choice unaddressed i e it could just be a code style cue to make a lambda more readable i ve primarily used across with mutate so the rationale in the summarise style example wasn t immediately obvious my suggestion would be to emphasise that a separate output is produced for each function e g a maybe clumsy alternative wording would be functions to apply to each of the selected columns where each transformation produces a separate output column based on the column and list names again i don t know for sure if this conflicts with any current or planned uses for the function as far as i can tell across more or less enforces this as a constraint i e requires unique names for each output although of course transform functions can return null in order to not produce a new output column optionally producing output for a list of functions opens up some interesting variations i guess that vary a little from the n x m account mostly my sense was that the intention or purpose behind the argument was underspecified in the documentation making it harder to understand the behaviour unless you already know the behaviour partly this could be a consequence of across being a fairly flexible tool so its meaning can look a little different in different uses but i think we can nudge the docs towards giving a quicker birds eye view hopefully you can see where i m coming from
| 0
|
5,280
| 8,069,066,686
|
IssuesEvent
|
2018-08-06 03:07:49
|
rubberduck-vba/Rubberduck
|
https://api.github.com/repos/rubberduck-vba/Rubberduck
|
opened
|
COM Collector should handle library imports better
|
discussion resolver typeinfo-processing
|
I was reminded by [a discussion in chat](https://chat.stackexchange.com/transcript/message/46026350#46026350) that I needed to look into 4 log items that pop up in almost every log file from Excel:
>
> 2018-08-05 21:20:45.4594;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBE;
> 2018-08-05 21:20:45.5341;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBE;
> 2018-08-05 21:20:45.8763;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBProject;
> 2018-08-05 21:20:45.9464;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBProject;
I had a suspicion about what was causing this, and just confirmed it - Excel has a reference to the VBA extensibility objects (which is obvious in that it's a host), and uses strongly typed members from that library on its interfaces, e.g.:
**Application.VBE**
```
[id(0x000004bb), propget, helpcontext(0x0002086b)]
VBE* VBE();
```
**Workbook.VBProject**
```
[id(0x000005bd), propget, helpcontext(0x00030a0d)]
VBProject* VBProject();
```
The COM Collector is resolving the `AsTypeName` just fine, because it will load the appropriate type library to resolve it:
```xml
<Node>
<Accessibility>Global</Accessibility>
<AsTypeName>VBProject</AsTypeName>
<Attributes />
<ComponentName>_Workbook</ComponentName>
<DeclarationType>PropertyGet</DeclarationType>
<DefaultValue i:nil="true" />
<Expression i:nil="true" />
<IdentifierName>VBProject</IdentifierName>
<IsArray>false</IsArray>
<IsByRefParam>false</IsByRefParam>
<IsControl>false</IsControl>
<IsExtensible>false</IsExtensible>
<IsOptionalParam>false</IsOptionalParam>
<IsParamArray>false</IsParamArray>
<IsSelfAssigned>false</IsSelfAssigned>
<IsUserDefined>false</IsUserDefined>
<IsWithEvents>false</IsWithEvents>
<MemberName>VBProject</MemberName>
<ParentScope>EXCEL.EXE;Excel._Workbook</ParentScope>
<ProjectName>Excel</ProjectName>
<ProjectPath>C:\Program Files\Microsoft Office\Office15\EXCEL.EXE</ProjectPath>
<TypeHint i:nil="true" />
</Node>
```
I'm not *entirely* sure how RD should handle these, in that it has more information available at compile time than the VBE does (the Object Browser also fails to browse them if they aren't referenced) :
```vba
Sub foo()
Dim project As VBProject '<-- Compile error: User-defined type not defined
Set project = ActiveWorkbook.VBProject
End Sub
```
On the one hand, without the reference, the VBE treats them as `Object` because it has to late bind them. This is problematic in the current collector because it only looks at a single reference - so if Project A has a reference to VBA extensibility and Project B doesn't, there really isn't a simple way to treat them differently.
On the other hand, this is really good information to know, because RD could use it as the basis for inspections like:
> Project Foo late binds VBProject, which is referenced by Excel. Consider adding a reference to Microsoft Visual Basic for Applications Extensibility.
I don't think there is a downside to the current handling (other than some log spam), but this could certainly be more robust. Thoughts?
In theory, we could do this for *any* forward reference by simply retrieving the name (for registered libs) or the path (for unregistered libs). We could even go a step further and pull in "unreferenced declarations" for the entire library so RD could inspect member accesses on the property *even if it is declared* `As Object`.
|
1.0
|
COM Collector should handle library imports better - I was reminded by [a discussion in chat](https://chat.stackexchange.com/transcript/message/46026350#46026350) that I needed to look into 4 log items that pop up in almost every log file from Excel:
>
> 2018-08-05 21:20:45.4594;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBE;
> 2018-08-05 21:20:45.5341;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBE;
> 2018-08-05 21:20:45.8763;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBProject;
> 2018-08-05 21:20:45.9464;WARN-2.2.6791.38398;Rubberduck.Parsing.Symbols.TypeAnnotationPass;Failed to resolve type VBProject;
I had a suspicion about what was causing this, and just confirmed it - Excel has a reference to the VBA extensibility objects (which is obvious in that it's a host), and uses strongly typed members from that library on its interfaces, e.g.:
**Application.VBE**
```
[id(0x000004bb), propget, helpcontext(0x0002086b)]
VBE* VBE();
```
**Workbook.VBProject**
```
[id(0x000005bd), propget, helpcontext(0x00030a0d)]
VBProject* VBProject();
```
The COM Collector is resolving the `AsTypeName` just fine, because it will load the appropriate type library to resolve it:
```xml
<Node>
<Accessibility>Global</Accessibility>
<AsTypeName>VBProject</AsTypeName>
<Attributes />
<ComponentName>_Workbook</ComponentName>
<DeclarationType>PropertyGet</DeclarationType>
<DefaultValue i:nil="true" />
<Expression i:nil="true" />
<IdentifierName>VBProject</IdentifierName>
<IsArray>false</IsArray>
<IsByRefParam>false</IsByRefParam>
<IsControl>false</IsControl>
<IsExtensible>false</IsExtensible>
<IsOptionalParam>false</IsOptionalParam>
<IsParamArray>false</IsParamArray>
<IsSelfAssigned>false</IsSelfAssigned>
<IsUserDefined>false</IsUserDefined>
<IsWithEvents>false</IsWithEvents>
<MemberName>VBProject</MemberName>
<ParentScope>EXCEL.EXE;Excel._Workbook</ParentScope>
<ProjectName>Excel</ProjectName>
<ProjectPath>C:\Program Files\Microsoft Office\Office15\EXCEL.EXE</ProjectPath>
<TypeHint i:nil="true" />
</Node>
```
I'm not *entirely* sure how RD should handle these, in that it has more information available at compile time than the VBE does (the Object Browser also fails to browse them if they aren't referenced) :
```vba
Sub foo()
Dim project As VBProject '<-- Compile error: User-defined type not defined
Set project = ActiveWorkbook.VBProject
End Sub
```
On the one hand, without the reference, the VBE treats them as `Object` because it has to late bind them. This is problematic in the current collector because it only looks at a single reference - so if Project A has a reference to VBA extensibility and Project B doesn't, there really isn't a simple way to treat them differently.
On the other hand, this is really good information to know, because RD could use it as the basis for inspections like:
> Project Foo late binds VBProject, which is referenced by Excel. Consider adding a reference to Microsoft Visual Basic for Applications Extensibility.
I don't think there is a downside to the current handling (other than some log spam), but this could certainly be more robust. Thoughts?
In theory, we could do this for *any* forward reference by simply retrieving the name (for registered libs) or the path (for unregistered libs). We could even go a step further and pull in "unreferenced declarations" for the entire library so RD could inspect member accesses on the property *even if it is declared* `As Object`.
|
process
|
com collector should handle library imports better i was reminded by that i needed to look into log items that pop up in almost every log file from excel warn rubberduck parsing symbols typeannotationpass failed to resolve type vbe warn rubberduck parsing symbols typeannotationpass failed to resolve type vbe warn rubberduck parsing symbols typeannotationpass failed to resolve type vbproject warn rubberduck parsing symbols typeannotationpass failed to resolve type vbproject i had a suspicion about what was causing this and just confirmed it excel has a reference to the vba extensibility objects which is obvious in that it s a host and uses strongly typed members from that library on its interfaces e g application vbe vbe vbe workbook vbproject vbproject vbproject the com collector is resolving the astypename just fine because it will load the appropriate type library to resolve it xml global vbproject workbook propertyget vbproject false false false false false false false false false vbproject excel exe excel workbook excel c program files microsoft office excel exe i m not entirely sure how rd should handle these in that it has more information available at compile time than the vbe does the object browser also fails to browse them if they aren t referenced vba sub foo dim project as vbproject compile error user defined type not defined set project activeworkbook vbproject end sub on the one hand without the reference the vbe treats them as object because it has to late bind them this is problematic in the current collector because it only looks at a single reference so if project a has a reference to vba extensibility and project b doesn t there really isn t a simple way to treat them differently on the other hand this is really good information to know because rd could use it as the basis for inspections like project foo late binds vbproject which is referenced by excel consider adding a reference to microsoft visual basic for applications extensibility i don t think there is a downside to the current handling other than some log spam but this could certainly be more robust thoughts in theory we could do this for any forward reference by simply retrieving the name for registered libs or the path for unregistered libs we could even go a step further and pull in unreferenced declarations for the entire library so rd could inspect member accesses on the property even if it is declared as object
| 1
|
31,904
| 8,773,804,829
|
IssuesEvent
|
2018-12-18 17:53:53
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
Azure managed image builder broken in 1.3.3 when snapshots are not enabled
|
bug builder/azure regression
|
Seems like 1.3.3 has a regression here: we have an Azure packer build (using the `hashicorp/packer:latest` Docker image) that suddenly started failing with the following error message with no changes other than the 1.3.3 upgrade:
```
==> azure-arm: Taking snapshot of OS disk ...
==> azure-arm: ERROR: -> UnsupportedResourceOperation : The resource type 'snapshots' does not support this operation.
```
I suspect this may be related to this feature from the 1.3.3 release:
> builder/azure: Add options for Managed Image OS Disk and Data Disk snapshots [GH-6980]
However, we're not using the new `managed_image_os_disk_snapshot_name` / `managed_image_data_disk_snapshot_prefix` options.
I will work on getting you guys the debug log output and a repro case - just thought you might want to know about this issue sooner rather than later.
|
1.0
|
Azure managed image builder broken in 1.3.3 when snapshots are not enabled - Seems like 1.3.3 has a regression here: we have an Azure packer build (using the `hashicorp/packer:latest` Docker image) that suddenly started failing with the following error message with no changes other than the 1.3.3 upgrade:
```
==> azure-arm: Taking snapshot of OS disk ...
==> azure-arm: ERROR: -> UnsupportedResourceOperation : The resource type 'snapshots' does not support this operation.
```
I suspect this may be related to this feature from the 1.3.3 release:
> builder/azure: Add options for Managed Image OS Disk and Data Disk snapshots [GH-6980]
However, we're not using the new `managed_image_os_disk_snapshot_name` / `managed_image_data_disk_snapshot_prefix` options.
I will work on getting you guys the debug log output and a repro case - just thought you might want to know about this issue sooner rather than later.
|
non_process
|
azure managed image builder broken in when snapshots are not enabled seems like has a regression here we have an azure packer build using the hashicorp packer latest docker image that suddenly started failing with the following error message with no changes other than the upgrade azure arm taking snapshot of os disk azure arm error unsupportedresourceoperation the resource type snapshots does not support this operation i suspect this may be related to this feature from the release builder azure add options for managed image os disk and data disk snapshots however we re not using the new managed image os disk snapshot name managed image data disk snapshot prefix options i will work on getting you guys the debug log output and a repro case just thought you might want to know about this issue sooner rather than later
| 0
|
21,740
| 30,257,511,882
|
IssuesEvent
|
2023-07-07 04:56:31
|
Significant-Gravitas/Auto-GPT
|
https://api.github.com/repos/Significant-Gravitas/Auto-GPT
|
closed
|
Command browse_website returned: Error: 'str' object has no attribute 'fast_llm_model'
|
function: process text
|
### β οΈ Search for existing issues first β οΈ
- [X] I have searched the existing issues, and there is no existing issue for my problem
### Which Operating System are you using?
Windows
### Which version of Auto-GPT are you using?
Latest Release
### Do you use OpenAI GPT-3 or GPT-4?
GPT-3.5
### Which area covers your issue best?
Commands
### Describe your issue.
browse_website command pulls information, gives errors in logs. This is different from [This issue](https://github.com/Significant-Gravitas/Auto-GPT/issues/4776)
Command browse_website returned: Error: 'str' object has no attribute 'fast_llm_model'
Please see [Logs here](https://autogpt.e2b.dev/) under the same name as this issue.
### Upload Activity Log Content
_No response_
### Upload Error Log Content
_No response_
|
1.0
|
Command browse_website returned: Error: 'str' object has no attribute 'fast_llm_model' - ### β οΈ Search for existing issues first β οΈ
- [X] I have searched the existing issues, and there is no existing issue for my problem
### Which Operating System are you using?
Windows
### Which version of Auto-GPT are you using?
Latest Release
### Do you use OpenAI GPT-3 or GPT-4?
GPT-3.5
### Which area covers your issue best?
Commands
### Describe your issue.
browse_website command pulls information, gives errors in logs. This is different from [This issue](https://github.com/Significant-Gravitas/Auto-GPT/issues/4776)
Command browse_website returned: Error: 'str' object has no attribute 'fast_llm_model'
Please see [Logs here](https://autogpt.e2b.dev/) under the same name as this issue.
### Upload Activity Log Content
_No response_
### Upload Error Log Content
_No response_
|
process
|
command browse website returned error str object has no attribute fast llm model β οΈ search for existing issues first β οΈ i have searched the existing issues and there is no existing issue for my problem which operating system are you using windows which version of auto gpt are you using latest release do you use openai gpt or gpt gpt which area covers your issue best commands describe your issue browse website command pulls information gives errors in logs this is different from command browse website returned error str object has no attribute fast llm model please see under the same name as this issue upload activity log content no response upload error log content no response
| 1
|
207,092
| 23,421,994,543
|
IssuesEvent
|
2022-08-13 21:03:32
|
nexmo-community/2fa-workflows
|
https://api.github.com/repos/nexmo-community/2fa-workflows
|
opened
|
connect-sqlite3-0.9.11.tgz: 1 vulnerabilities (highest severity is: 7.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>connect-sqlite3-0.9.11.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/connect-sqlite3/node_modules/sqlite3/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/2fa-workflows/commit/a207b34cc844a5f67eee17cbce2edb80b3012d1e">a207b34cc844a5f67eee17cbce2edb80b3012d1e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-21227](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21227) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | sqlite3-4.2.0.tgz | Transitive | 0.9.12 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-21227</summary>
### Vulnerable Library - <b>sqlite3-4.2.0.tgz</b></p>
<p>Asynchronous, non-blocking SQLite3 bindings</p>
<p>Library home page: <a href="https://registry.npmjs.org/sqlite3/-/sqlite3-4.2.0.tgz">https://registry.npmjs.org/sqlite3/-/sqlite3-4.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/connect-sqlite3/node_modules/sqlite3/package.json</p>
<p>
Dependency Hierarchy:
- connect-sqlite3-0.9.11.tgz (Root Library)
- :x: **sqlite3-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/2fa-workflows/commit/a207b34cc844a5f67eee17cbce2edb80b3012d1e">a207b34cc844a5f67eee17cbce2edb80b3012d1e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package sqlite3 before 5.0.3 are vulnerable to Denial of Service (DoS) which will invoke the toString function of the passed parameter. If passed an invalid Function object it will throw and crash the V8 engine.
<p>Publish Date: 2022-05-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21227>CVE-2022-21227</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9qrh-qjmc-5w2p">https://github.com/advisories/GHSA-9qrh-qjmc-5w2p</a></p>
<p>Release Date: 2022-05-01</p>
<p>Fix Resolution (sqlite3): 5.0.3</p>
<p>Direct dependency fix Resolution (connect-sqlite3): 0.9.12</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
True
|
connect-sqlite3-0.9.11.tgz: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>connect-sqlite3-0.9.11.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/connect-sqlite3/node_modules/sqlite3/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/2fa-workflows/commit/a207b34cc844a5f67eee17cbce2edb80b3012d1e">a207b34cc844a5f67eee17cbce2edb80b3012d1e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-21227](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21227) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | sqlite3-4.2.0.tgz | Transitive | 0.9.12 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-21227</summary>
### Vulnerable Library - <b>sqlite3-4.2.0.tgz</b></p>
<p>Asynchronous, non-blocking SQLite3 bindings</p>
<p>Library home page: <a href="https://registry.npmjs.org/sqlite3/-/sqlite3-4.2.0.tgz">https://registry.npmjs.org/sqlite3/-/sqlite3-4.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/connect-sqlite3/node_modules/sqlite3/package.json</p>
<p>
Dependency Hierarchy:
- connect-sqlite3-0.9.11.tgz (Root Library)
- :x: **sqlite3-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nexmo-community/2fa-workflows/commit/a207b34cc844a5f67eee17cbce2edb80b3012d1e">a207b34cc844a5f67eee17cbce2edb80b3012d1e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package sqlite3 before 5.0.3 are vulnerable to Denial of Service (DoS) which will invoke the toString function of the passed parameter. If passed an invalid Function object it will throw and crash the V8 engine.
<p>Publish Date: 2022-05-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21227>CVE-2022-21227</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9qrh-qjmc-5w2p">https://github.com/advisories/GHSA-9qrh-qjmc-5w2p</a></p>
<p>Release Date: 2022-05-01</p>
<p>Fix Resolution (sqlite3): 5.0.3</p>
<p>Direct dependency fix Resolution (connect-sqlite3): 0.9.12</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
non_process
|
connect tgz vulnerabilities highest severity is vulnerable library connect tgz path to dependency file package json path to vulnerable library node modules connect node modules package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high tgz transitive details cve vulnerable library tgz asynchronous non blocking bindings library home page a href path to dependency file package json path to vulnerable library node modules connect node modules package json dependency hierarchy connect tgz root library x tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package before are vulnerable to denial of service dos which will invoke the tostring function of the passed parameter if passed an invalid function object it will throw and crash the engine publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution direct dependency fix resolution connect rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
| 0
|
51,262
| 6,506,728,550
|
IssuesEvent
|
2017-08-24 10:10:20
|
owncloud/client
|
https://api.github.com/repos/owncloud/client
|
closed
|
Newer file gets renamed to _conflict, should be the other way around
|
Design & UX Discussion
|
The current state is this:

The newer (top) file which also is the correct one gets the _conflict status. However, in most of these cases the more recent file is the correct one, and the older one should get the _conflict status. Making the recent one to look like itβs _only_ a conflict is confusing.
Currently to fix the conflict, I need to:
1. Check and really make sure that _conflict is indeed the correct one
2. Delete the other file
3. Rename the Β»conflictΒ« file to remove the Β»_conflictΒ« part from the name
If it would be the other way around, I could just simply quickly look at it, see that the conflict file is older anyway, and delete that one.
cc @ckamm @ogoffart @guruz what do you think? I am continuously confused by this.
|
1.0
|
Newer file gets renamed to _conflict, should be the other way around - The current state is this:

The newer (top) file which also is the correct one gets the _conflict status. However, in most of these cases the more recent file is the correct one, and the older one should get the _conflict status. Making the recent one to look like itβs _only_ a conflict is confusing.
Currently to fix the conflict, I need to:
1. Check and really make sure that _conflict is indeed the correct one
2. Delete the other file
3. Rename the Β»conflictΒ« file to remove the Β»_conflictΒ« part from the name
If it would be the other way around, I could just simply quickly look at it, see that the conflict file is older anyway, and delete that one.
cc @ckamm @ogoffart @guruz what do you think? I am continuously confused by this.
|
non_process
|
newer file gets renamed to conflict should be the other way around the current state is this the newer top file which also is the correct one gets the conflict status however in most of these cases the more recent file is the correct one and the older one should get the conflict status making the recent one to look like itβs only a conflict is confusing currently to fix the conflict i need to check and really make sure that conflict is indeed the correct one delete the other file rename the Β»conflictΒ« file to remove the Β» conflictΒ« part from the name if it would be the other way around i could just simply quickly look at it see that the conflict file is older anyway and delete that one cc ckamm ogoffart guruz what do you think i am continuously confused by this
| 0
|
762,547
| 26,722,620,543
|
IssuesEvent
|
2023-01-29 10:20:44
|
KienVu1504/SuperCook_FE
|
https://api.github.com/repos/KienVu1504/SuperCook_FE
|
closed
|
ingredient categories manage page
|
Done Request Priority Admin
|
**Describe the solution you'd like**
1. add ingredient categories manage page (list table with remove&create&edit action)
2. form input on modal
**Additional context**

|
1.0
|
ingredient categories manage page -
**Describe the solution you'd like**
1. add ingredient categories manage page (list table with remove&create&edit action)
2. form input on modal
**Additional context**

|
non_process
|
ingredient categories manage page describe the solution you d like add ingredient categories manage page list table with remove create edit action form input on modal additional context
| 0
|
18,988
| 24,979,543,304
|
IssuesEvent
|
2022-11-02 10:35:14
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Create sibling to 'viral release by cytolysis via suppression of host peptidoglycan biosynthetic process'
|
New term request multi-species process
|
to describe
'viral release by cytolysis via disruption of host peptidoglycan cell wall'
|
1.0
|
Create sibling to 'viral release by cytolysis via suppression of host peptidoglycan biosynthetic process' - to describe
'viral release by cytolysis via disruption of host peptidoglycan cell wall'
|
process
|
create sibling to viral release by cytolysis via suppression of host peptidoglycan biosynthetic process to describe viral release by cytolysis via disruption of host peptidoglycan cell wall
| 1
|
371,084
| 10,961,238,967
|
IssuesEvent
|
2019-11-27 15:02:46
|
inverse-inc/packetfence
|
https://api.github.com/repos/inverse-inc/packetfence
|
closed
|
In multi_zone config the db server list is not ordered by the position in cluster.conf
|
Priority: High Type: Bug
|
**Describe the bug**
Multizone cluster config, 5 DC, the cluster.conf is the same on each servers.
When you generate haproxy-db configuration, it doesn't respect the order in the file.
|
1.0
|
In multi_zone config the db server list is not ordered by the position in cluster.conf - **Describe the bug**
Multizone cluster config, 5 DC, the cluster.conf is the same on each servers.
When you generate haproxy-db configuration, it doesn't respect the order in the file.
|
non_process
|
in multi zone config the db server list is not ordered by the position in cluster conf describe the bug multizone cluster config dc the cluster conf is the same on each servers when you generate haproxy db configuration it doesn t respect the order in the file
| 0
|
7,510
| 10,589,292,776
|
IssuesEvent
|
2019-10-09 05:37:02
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Documentation should be rewritten to use Azure PowerShell Az
|
Pri2 automation/svc cxp process-automation/subsvc triaged
|
This documentation is using AzureRM when new users should use Azure PowerShell Az. I think this documentation should be rewritten. :)
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: cdf26281-619d-1997-1a7f-49ef9f0e261c
* Version Independent ID: c0f4d49e-a0a3-d8b7-c713-5800147a452b
* Content: [My first PowerShell Workflow runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-first-runbook-textual#feedback)
* Content Source: [articles/automation/automation-first-runbook-textual.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-first-runbook-textual.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @bobbytreed
* Microsoft Alias: **robreed**
|
1.0
|
Documentation should be rewritten to use Azure PowerShell Az - This documentation is using AzureRM when new users should use Azure PowerShell Az. I think this documentation should be rewritten. :)
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: cdf26281-619d-1997-1a7f-49ef9f0e261c
* Version Independent ID: c0f4d49e-a0a3-d8b7-c713-5800147a452b
* Content: [My first PowerShell Workflow runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-first-runbook-textual#feedback)
* Content Source: [articles/automation/automation-first-runbook-textual.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-first-runbook-textual.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @bobbytreed
* Microsoft Alias: **robreed**
|
process
|
documentation should be rewritten to use azure powershell az this documentation is using azurerm when new users should use azure powershell az i think this documentation should be rewritten document details β do not edit this section it is required for docs microsoft com β github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
| 1
|
60,182
| 25,023,919,904
|
IssuesEvent
|
2022-11-04 05:23:11
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
confuse about Isolated and I1v2
|
app-service/svc triaged cxp product-question Pri1
|

Does **Isolated** in the above picture refer to creating new app service environment v1?

If so, does IsolatedV2 refer to creating new app service environment v2οΌ
What does I1v2 and I2v1 respectively refer to?

Please help to confirm. Thanks a lot.
---
#### Document Details
β *Do not edit this section. It is required for learn.microsoft.com β GitHub issue linking.*
* ID: 09a9696e-10fd-a74e-4957-0d7ed9ef5a0f
* Version Independent ID: dad3cbd5-b2d0-9cc4-261f-97b95cf811c6
* Content: [App Service plans - Azure App Service](https://learn.microsoft.com/en-us/azure/app-service/overview-hosting-plans)
* Content Source: [articles/app-service/overview-hosting-plans.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/app-service/overview-hosting-plans.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin**
|
1.0
|
confuse about Isolated and I1v2 -

Does **Isolated** in the above picture refer to creating new app service environment v1?

If so, does IsolatedV2 refer to creating new app service environment v2οΌ
What does I1v2 and I2v1 respectively refer to?

Please help to confirm. Thanks a lot.
---
#### Document Details
β *Do not edit this section. It is required for learn.microsoft.com β GitHub issue linking.*
* ID: 09a9696e-10fd-a74e-4957-0d7ed9ef5a0f
* Version Independent ID: dad3cbd5-b2d0-9cc4-261f-97b95cf811c6
* Content: [App Service plans - Azure App Service](https://learn.microsoft.com/en-us/azure/app-service/overview-hosting-plans)
* Content Source: [articles/app-service/overview-hosting-plans.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/app-service/overview-hosting-plans.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin**
|
non_process
|
confuse about isolated and does isolated in the above picture refer to creating new app service environment if so does refer to creating new app service environment οΌ what does and respectively refer to please help to confirm thanks a lot document details β do not edit this section it is required for learn microsoft com β github issue linking id version independent id content content source service app service github login cephalin microsoft alias cephalin
| 0
|
5,093
| 7,878,570,692
|
IssuesEvent
|
2018-06-26 10:39:50
|
RustyPanda/zoobot
|
https://api.github.com/repos/RustyPanda/zoobot
|
opened
|
Add astro-suitable normalization of image
|
Preprocessing enhancement
|
Common practice is to subtract off the mean image. I can experiment with this, but I'm not sure how well it applies in astro context.
The tf function fails:
image = tf.image.per_image_standardization(image)
|
1.0
|
Add astro-suitable normalization of image - Common practice is to subtract off the mean image. I can experiment with this, but I'm not sure how well it applies in astro context.
The tf function fails:
image = tf.image.per_image_standardization(image)
|
process
|
add astro suitable normalization of image common practice is to subtract off the mean image i can experiment with this but i m not sure how well it applies in astro context the tf function fails image tf image per image standardization image
| 1
|
3,612
| 6,653,507,948
|
IssuesEvent
|
2017-09-29 08:42:12
|
bazelbuild/continuous-integration
|
https://api.github.com/repos/bazelbuild/continuous-integration
|
opened
|
Identify and train a support team for ci.bazel.io
|
P1 process
|
Rather than have myself be the relay for all the issue for both sheriff and users of the system.
|
1.0
|
Identify and train a support team for ci.bazel.io - Rather than have myself be the relay for all the issue for both sheriff and users of the system.
|
process
|
identify and train a support team for ci bazel io rather than have myself be the relay for all the issue for both sheriff and users of the system
| 1
|
14,724
| 17,936,294,903
|
IssuesEvent
|
2021-09-10 15:43:50
|
bridgetownrb/bridgetown
|
https://api.github.com/repos/bridgetownrb/bridgetown
|
opened
|
Refactor the pagination gem once Resource engine is default
|
process resource engine
|
The pagination gem largely contains code ported from a third-party gem written for Jekyll, and even though I've made a number of enhancements/performance fixes since adding it in, a lot of it is still pretty crufty and not very Ruby-like IMHO (aka lots of linear procedural-style code and very long methods).
Once the Resource engine reigns supreme, I'd like to start pretty much from scratch and ensure the pagination subsystem is well-written, performant, and easily hooks into new Resource features like taxonomies. There are also some funky things around how it current handles permalink and file path outputs we could clean up.
|
1.0
|
Refactor the pagination gem once Resource engine is default - The pagination gem largely contains code ported from a third-party gem written for Jekyll, and even though I've made a number of enhancements/performance fixes since adding it in, a lot of it is still pretty crufty and not very Ruby-like IMHO (aka lots of linear procedural-style code and very long methods).
Once the Resource engine reigns supreme, I'd like to start pretty much from scratch and ensure the pagination subsystem is well-written, performant, and easily hooks into new Resource features like taxonomies. There are also some funky things around how it current handles permalink and file path outputs we could clean up.
|
process
|
refactor the pagination gem once resource engine is default the pagination gem largely contains code ported from a third party gem written for jekyll and even though i ve made a number of enhancements performance fixes since adding it in a lot of it is still pretty crufty and not very ruby like imho aka lots of linear procedural style code and very long methods once the resource engine reigns supreme i d like to start pretty much from scratch and ensure the pagination subsystem is well written performant and easily hooks into new resource features like taxonomies there are also some funky things around how it current handles permalink and file path outputs we could clean up
| 1
|
13,260
| 15,729,101,509
|
IssuesEvent
|
2021-03-29 14:30:33
|
esmero/strawberry_runners
|
https://api.github.com/repos/esmero/strawberry_runners
|
closed
|
Add plaintext and Total Sequence Count to Search API indexable OCR processor
|
Datapackage / Frictionless Post processor Plugins Solr Indexing enhancement
|
# What is this?
Matching issue for https://github.com/esmero/strawberryfield/pull/168 and https://github.com/esmero/strawberryfield/issue/165
This will make HOCR processor pass 2 new elements back to the Abstract Processor allowing pure, plain text and an expected total count of documents to be indexed in Solr. The first one is needed for nice Search Excerpts, the second to allow a "harvesting when ready" and saving back into a Frictionless data package at ADO level for long time persistence of generated HOCR. (expensive stuff to generate every time).
|
1.0
|
Add plaintext and Total Sequence Count to Search API indexable OCR processor - # What is this?
Matching issue for https://github.com/esmero/strawberryfield/pull/168 and https://github.com/esmero/strawberryfield/issue/165
This will make HOCR processor pass 2 new elements back to the Abstract Processor allowing pure, plain text and an expected total count of documents to be indexed in Solr. The first one is needed for nice Search Excerpts, the second to allow a "harvesting when ready" and saving back into a Frictionless data package at ADO level for long time persistence of generated HOCR. (expensive stuff to generate every time).
|
process
|
add plaintext and total sequence count to search api indexable ocr processor what is this matching issue for and this will make hocr processor pass new elements back to the abstract processor allowing pure plain text and an expected total count of documents to be indexed in solr the first one is needed for nice search excerpts the second to allow a harvesting when ready and saving back into a frictionless data package at ado level for long time persistence of generated hocr expensive stuff to generate every time
| 1
|
14,820
| 18,157,366,078
|
IssuesEvent
|
2021-09-27 04:40:45
|
dotnetcore/Home
|
https://api.github.com/repos/dotnetcore/Home
|
closed
|
BigCookieKit η³θ―·ε ε
₯
|
Ap: Process-Termination Np: Application
|
## Basic
Project Name:BigCookieKit
Project Address on GitHub (**and** other addresses, such as Gitee or GitLab):https://github.com/BigBigZBBing/BigCookieKit
Project Description:ι«ζ§θ½εε° Officeι«ζ§θ½θ―»ε
Document, Blog or Wiki address:https://blog.csdn.net/weixin_42995482
Author:Big.Cookie
Development team or main contributors:Big.Cookie
License:MIT
## Additional information
- [x] 1 - The project is based on .NET technology.
- [x] 2 - The project has a clear git commit log.
- [x] 3 - Unit tests with considerable coverage
- [x] 4 - The project has benchmarking information (for infrastructure projects, this paragraph should be met)
- [x] 5 - The project creation time should be at least three natural months from the application time
- [x] 6 - The core developers of the project team should make effective contributions to the project within one natural month from the application time
- [x] 7 - Project should be hosted on GitHub first
- [ ] 8 - The number of stars in the project is not less than 100
- [x] 9 - Project should have more complete information:
- [x] Β§9.1 - README (, with internationalization is better)
- [ ] Β§9.2 - Documentation or Wiki
- [x] Β§9.3 - Sample code
- [ ] Β§9.4 - Roadmap
- [ ] Β§9.5 - Other options: website, blog, manual, tutorial or publication
- [x] 10 - The project should have more reliable technical support and response capabilities:
- [x] Β§10.1 - More effective issue response
- [x] Β§10.2 - Other options: communities, mail groups, groups and other social media channels, etc.
- [ ] 11 - The project has a clear development plan and roadmap
- [x] 12 - The project has not received sponsorship from commercial companies or organizations, and no company has paid for the project
- [x] 13 - The project has no copyright issues and meets the copyright requirements in the "Community Project Copyright and Open Source License Regulations";
|
1.0
|
BigCookieKit η³θ―·ε ε
₯ - ## Basic
Project Name:BigCookieKit
Project Address on GitHub (**and** other addresses, such as Gitee or GitLab):https://github.com/BigBigZBBing/BigCookieKit
Project Description:ι«ζ§θ½εε° Officeι«ζ§θ½θ―»ε
Document, Blog or Wiki address:https://blog.csdn.net/weixin_42995482
Author:Big.Cookie
Development team or main contributors:Big.Cookie
License:MIT
## Additional information
- [x] 1 - The project is based on .NET technology.
- [x] 2 - The project has a clear git commit log.
- [x] 3 - Unit tests with considerable coverage
- [x] 4 - The project has benchmarking information (for infrastructure projects, this paragraph should be met)
- [x] 5 - The project creation time should be at least three natural months from the application time
- [x] 6 - The core developers of the project team should make effective contributions to the project within one natural month from the application time
- [x] 7 - Project should be hosted on GitHub first
- [ ] 8 - The number of stars in the project is not less than 100
- [x] 9 - Project should have more complete information:
- [x] Β§9.1 - README (, with internationalization is better)
- [ ] Β§9.2 - Documentation or Wiki
- [x] Β§9.3 - Sample code
- [ ] Β§9.4 - Roadmap
- [ ] Β§9.5 - Other options: website, blog, manual, tutorial or publication
- [x] 10 - The project should have more reliable technical support and response capabilities:
- [x] Β§10.1 - More effective issue response
- [x] Β§10.2 - Other options: communities, mail groups, groups and other social media channels, etc.
- [ ] 11 - The project has a clear development plan and roadmap
- [x] 12 - The project has not received sponsorship from commercial companies or organizations, and no company has paid for the project
- [x] 13 - The project has no copyright issues and meets the copyright requirements in the "Community Project Copyright and Open Source License Regulations";
|
process
|
bigcookiekit η³θ―·ε ε
₯ basic project name bigcookiekit project address on github and other addresses such as gitee or gitlab project description ι«ζ§θ½εε° officeι«ζ§θ½θ―»ε document blog or wiki address author big cookie development team or main contributors big cookie license mit additional information the project is based on net technology the project has a clear git commit log unit tests with considerable coverage the project has benchmarking information for infrastructure projects this paragraph should be met the project creation time should be at least three natural months from the application time the core developers of the project team should make effective contributions to the project within one natural month from the application time project should be hosted on github first the number of stars in the project is not less than project should have more complete information Β§ readme with internationalization is better Β§ documentation or wiki Β§ sample code Β§ roadmap Β§ other options website blog manual tutorial or publication the project should have more reliable technical support and response capabilities Β§ more effective issue response Β§ other options communities mail groups groups and other social media channels etc the project has a clear development plan and roadmap the project has not received sponsorship from commercial companies or organizations and no company has paid for the project the project has no copyright issues and meets the copyright requirements in the community project copyright and open source license regulations
| 1
|
18,997
| 11,103,590,724
|
IssuesEvent
|
2019-12-17 04:32:08
|
gctools-outilsgc/gcconnex
|
https://api.github.com/repos/gctools-outilsgc/gcconnex
|
closed
|
Error when logging into GCcollab
|
Project: Account Service: gccollab [zube]: In Review bug
|
Some users are encountering an error page when attempting to login. Example user:
jm@publicsectorcompany.com
## Details on issue or enhancement

## For the development team
- [ ] Issue user story documented
- [ ] UX input received
- [ ] Design completed
- [ ] Design validated by business team / UX
- [ ] Code review completed by peer
- [ ] Issue closing comment references any duplicate or connected issues or pull requests
- [ ] Issue closed
|
1.0
|
Error when logging into GCcollab - Some users are encountering an error page when attempting to login. Example user:
jm@publicsectorcompany.com
## Details on issue or enhancement

## For the development team
- [ ] Issue user story documented
- [ ] UX input received
- [ ] Design completed
- [ ] Design validated by business team / UX
- [ ] Code review completed by peer
- [ ] Issue closing comment references any duplicate or connected issues or pull requests
- [ ] Issue closed
|
non_process
|
error when logging into gccollab some users are encountering an error page when attempting to login example user jm publicsectorcompany com details on issue or enhancement for the development team issue user story documented ux input received design completed design validated by business team ux code review completed by peer issue closing comment references any duplicate or connected issues or pull requests issue closed
| 0
|
33,139
| 14,006,631,062
|
IssuesEvent
|
2020-10-28 20:15:57
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
Weekly Scrum of Scrums Calls - ATD Backlog Development
|
Need: 2-Should Have Product: AMANDA Service: Product Type: Meeting Workgroup: DTS
|
### Tuesdays & Fridays @ 1:30pm with leads from each team (ATD, LIT, JETS)
|
1.0
|
Weekly Scrum of Scrums Calls - ATD Backlog Development - ### Tuesdays & Fridays @ 1:30pm with leads from each team (ATD, LIT, JETS)
|
non_process
|
weekly scrum of scrums calls atd backlog development tuesdays fridays with leads from each team atd lit jets
| 0
|
277,506
| 30,659,273,613
|
IssuesEvent
|
2023-07-25 14:04:22
|
rsoreq/zenbot
|
https://api.github.com/repos/rsoreq/zenbot
|
closed
|
CVE-2022-0122 (Medium) detected in node-forge-0.7.6.tgz - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2022-0122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.7.6.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz</a></p>
<p>
Dependency Hierarchy:
- pushbullet-2.4.0.tgz (Root Library)
- :x: **node-forge-0.7.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/zenbot/commit/7a24c0d7b98ee76e6bac827974cff490a7694378">7a24c0d7b98ee76e6bac827974cff490a7694378</a></p>
<p>Found in base branch: <b>unstable</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
forge is vulnerable to URL Redirection to Untrusted Site
Mend Note: Converted from WS-2022-0007, on 2022-11-07.
<p>Publish Date: 2022-01-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0122>CVE-2022-0122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p>
<p>Release Date: 2022-01-06</p>
<p>Fix Resolution (node-forge): 1.0.0</p>
<p>Direct dependency fix Resolution (pushbullet): 3.0.0</p>
</p>
</details>
<p></p>
|
True
|
CVE-2022-0122 (Medium) detected in node-forge-0.7.6.tgz - autoclosed - ## CVE-2022-0122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.7.6.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz</a></p>
<p>
Dependency Hierarchy:
- pushbullet-2.4.0.tgz (Root Library)
- :x: **node-forge-0.7.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/zenbot/commit/7a24c0d7b98ee76e6bac827974cff490a7694378">7a24c0d7b98ee76e6bac827974cff490a7694378</a></p>
<p>Found in base branch: <b>unstable</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
forge is vulnerable to URL Redirection to Untrusted Site
Mend Note: Converted from WS-2022-0007, on 2022-11-07.
<p>Publish Date: 2022-01-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0122>CVE-2022-0122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p>
<p>Release Date: 2022-01-06</p>
<p>Fix Resolution (node-forge): 1.0.0</p>
<p>Direct dependency fix Resolution (pushbullet): 3.0.0</p>
</p>
</details>
<p></p>
|
non_process
|
cve medium detected in node forge tgz autoclosed cve medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href dependency hierarchy pushbullet tgz root library x node forge tgz vulnerable library found in head commit a href found in base branch unstable vulnerability details forge is vulnerable to url redirection to untrusted site mend note converted from ws on publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution pushbullet
| 0
|
18,626
| 24,579,709,713
|
IssuesEvent
|
2022-10-13 14:46:58
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Android] [Consent API] Consent record with revoked state should be created ,when participant account is deleted in the mobile app
|
Bug P1 Android Response datastore Process: Fixed Process: Tested QA Process: Tested dev
|
**Pre-condition:** Study should be created in the study builder
**Steps:**
1. Sign in / Sign up
2. Enroll for the study
3. Now go to 'My account' section and delete the user account
4. Go to that particular consent record and Verify
**AR:** Consent record with revoked state is not getting created ,when participant account is deleted in the mobile app
**ER:** Consent record with revoked state should be created ,when participant account is deleted in the mobile app
**Note:** Issue is observed only when response is submitted from android 12 device
|
3.0
|
[Android] [Consent API] Consent record with revoked state should be created ,when participant account is deleted in the mobile app - **Pre-condition:** Study should be created in the study builder
**Steps:**
1. Sign in / Sign up
2. Enroll for the study
3. Now go to 'My account' section and delete the user account
4. Go to that particular consent record and Verify
**AR:** Consent record with revoked state is not getting created ,when participant account is deleted in the mobile app
**ER:** Consent record with revoked state should be created ,when participant account is deleted in the mobile app
**Note:** Issue is observed only when response is submitted from android 12 device
|
process
|
consent record with revoked state should be created when participant account is deleted in the mobile app pre condition study should be created in the study builder steps sign in sign up enroll for the study now go to my account section and delete the user account go to that particular consent record and verify ar consent record with revoked state is not getting created when participant account is deleted in the mobile app er consent record with revoked state should be created when participant account is deleted in the mobile app note issue is observed only when response is submitted from android device
| 1
|
14,607
| 11,010,835,310
|
IssuesEvent
|
2019-12-04 15:16:15
|
MeteoSwiss-APN/dawn
|
https://api.github.com/repos/MeteoSwiss-APN/dawn
|
closed
|
Add regression testing for the GT4py - Dawn chain
|
enhancement infrastructure sprint planning
|
in #408 we've created the interface between GT4py and dawn.
There should be regression tests that check this
|
1.0
|
Add regression testing for the GT4py - Dawn chain - in #408 we've created the interface between GT4py and dawn.
There should be regression tests that check this
|
non_process
|
add regression testing for the dawn chain in we ve created the interface between and dawn there should be regression tests that check this
| 0
|
1,867
| 4,697,420,966
|
IssuesEvent
|
2016-10-12 09:18:23
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child processes started with `--debug-brk` do not buffer messages
|
child_process debugger question
|
Child node process started with `--debug` receives messages sent before registering message listener in the child process.
However, child node processes started with `--debug-brk` don't receive messages.
Consider this example:
Initially, install & start node-inspector
```bash
$ npm install -g node-inspector && node-inspector
```
Create two files:
main.js
```javascript
function forkChild(debugOption) {
var child = require('child_process').fork('./child.js', [], {execArgv: [debugOption]});
console.log('child process started: ' + debugOption + ', pid: ' + child.pid);
child.send({say: 'hello to ' + debugOption});
console.log('message sent to child process started with ' + debugOption);
}
forkChild('--debug=5860');
forkChild('--debug-brk=5861');
```
child.js
```javascript
function wait(millis) {
var start = new Date().getTime();
while (new Date().getTime() < start + millis) {}
}
var prefix = '[' + require('path').basename(__filename) + ' ' + process.execArgv + ']';
console.log(prefix + ' started');
wait(3000);
process.on('message', function (msg) {
console.log(prefix + ' got message', msg);
});
```
Run main.js:
```
$ node main.js
child process started: --debug=5860, pid: 11808
message sent to child process started with --debug=5860
child process started: --debug-brk=5861, pid: 11814
message sent to child process started with --debug-brk=5861
Debugger listening on port 5860
[child.js --debug=5860] started
Debugger listening on port 5861
[child.js --debug=5860] got message { say: 'hello to --debug=5860' }
```
Visit `http://127.0.0.1:8080/?port=5861` (node-inspector should be already running) to resume the suspended execution of the second child process: node-inspector web page will open with execution suspended on the first line of child.js. Resume the execution (press F8 in Google Chrome). We'll get the following console output in the main process:
```
[child.js --debug-brk=5861] started
```
However, expected lines were:
```
[child.js --debug-brk=5861] started
[child.js --debug=5861] got message { say: 'hello to --debug-brk=5861' }
```
* **Version**: v5.9.0
* **Platform**: Linux home-computer 3.13.0-37-generic x64-Ubuntu SMP Mon Sep 22 21:28:38 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux
* **Subsystem**: child_process
|
1.0
|
child processes started with `--debug-brk` do not buffer messages - Child node process started with `--debug` receives messages sent before registering message listener in the child process.
However, child node processes started with `--debug-brk` don't receive messages.
Consider this example:
Initially, install & start node-inspector
```bash
$ npm install -g node-inspector && node-inspector
```
Create two files:
main.js
```javascript
function forkChild(debugOption) {
var child = require('child_process').fork('./child.js', [], {execArgv: [debugOption]});
console.log('child process started: ' + debugOption + ', pid: ' + child.pid);
child.send({say: 'hello to ' + debugOption});
console.log('message sent to child process started with ' + debugOption);
}
forkChild('--debug=5860');
forkChild('--debug-brk=5861');
```
child.js
```javascript
function wait(millis) {
var start = new Date().getTime();
while (new Date().getTime() < start + millis) {}
}
var prefix = '[' + require('path').basename(__filename) + ' ' + process.execArgv + ']';
console.log(prefix + ' started');
wait(3000);
process.on('message', function (msg) {
console.log(prefix + ' got message', msg);
});
```
Run main.js:
```
$ node main.js
child process started: --debug=5860, pid: 11808
message sent to child process started with --debug=5860
child process started: --debug-brk=5861, pid: 11814
message sent to child process started with --debug-brk=5861
Debugger listening on port 5860
[child.js --debug=5860] started
Debugger listening on port 5861
[child.js --debug=5860] got message { say: 'hello to --debug=5860' }
```
Visit `http://127.0.0.1:8080/?port=5861` (node-inspector should be already running) to resume the suspended execution of the second child process: node-inspector web page will open with execution suspended on the first line of child.js. Resume the execution (press F8 in Google Chrome). We'll get the following console output in the main process:
```
[child.js --debug-brk=5861] started
```
However, expected lines were:
```
[child.js --debug-brk=5861] started
[child.js --debug=5861] got message { say: 'hello to --debug-brk=5861' }
```
* **Version**: v5.9.0
* **Platform**: Linux home-computer 3.13.0-37-generic x64-Ubuntu SMP Mon Sep 22 21:28:38 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux
* **Subsystem**: child_process
|
process
|
child processes started with debug brk do not buffer messages child node process started with debug receives messages sent before registering message listener in the child process however child node processes started with debug brk don t receive messages consider this example initially install start node inspector bash npm install g node inspector node inspector create two files main js javascript function forkchild debugoption var child require child process fork child js execargv console log child process started debugoption pid child pid child send say hello to debugoption console log message sent to child process started with debugoption forkchild debug forkchild debug brk child js javascript function wait millis var start new date gettime while new date gettime start millis var prefix console log prefix started wait process on message function msg console log prefix got message msg run main js node main js child process started debug pid message sent to child process started with debug child process started debug brk pid message sent to child process started with debug brk debugger listening on port started debugger listening on port got message say hello to debug visit node inspector should be already running to resume the suspended execution of the second child process node inspector web page will open with execution suspended on the first line of child js resume the execution press in google chrome we ll get the following console output in the main process started however expected lines were started got message say hello to debug brk version platform linux home computer generic ubuntu smp mon sep utc gnu linux subsystem child process
| 1
|
178,936
| 14,685,562,204
|
IssuesEvent
|
2021-01-01 09:59:25
|
gramener/gramex-guide
|
https://api.github.com/repos/gramener/gramex-guide
|
closed
|
Document how to serve custom error message
|
documentation
|
raised by @manojraju-g
upon redirection from http to https, nginx serves a "301 moved permanently" message. This needs to be customized to not reveal the server name.
|
1.0
|
Document how to serve custom error message - raised by @manojraju-g
upon redirection from http to https, nginx serves a "301 moved permanently" message. This needs to be customized to not reveal the server name.
|
non_process
|
document how to serve custom error message raised by manojraju g upon redirection from http to https nginx serves a moved permanently message this needs to be customized to not reveal the server name
| 0
|
15,834
| 20,021,791,438
|
IssuesEvent
|
2022-02-01 17:02:43
|
varabyte/kobweb
|
https://api.github.com/repos/varabyte/kobweb
|
closed
|
Find better ways to distribute the kobweb binary
|
process
|
~~How to get onto package managers in Linux or Homebrew in Mac? What's the equivalent for Windows?~~
Support and/or intentionally reject the following package managers:
- [x] homebrew
- [x] scoop
- [X] ~~chocolatey~~ (1)
- [X] ~~gofish~~ (2)
- [X] ~~macports~~ (2)
- [X] ~~snapcraft~~ (3)
- [X] ~~spec~~ (4)
- [X] ~~docker~~ (5)
- [x] sdkman
* ~~strikethrough~~ means not now, we can close this bug without them, but I'd always be open to supporting them later
1. The process for Chocolatey feels a bit heavy handed for now. I can't simply self-host, instead I have to go through a whole process with their team, getting a review, etc. Plus, I found it hard to navigate the instructions. I'll leave this off for now, since Scoop is working, but can revisit this again in the future if Kobweb actually becomes regularly used.
2. Requires setting up PRs on specific repositories, not low hanging fruit
3. Strict mode is killing me. Snap sets up a whole environment just for Kobweb, making the download 100MB, and yet is getting killed by Snap and `snappy-debug` isn't really giving me useful information.
4. Not too familiar with Fedora or spec, but I'm considered adding support for it is diminishing returns and therefore isn't my highest priority.
5. A docker image is probably overkill. Users will be expected to have java on their machine anyway to compile Kobweb (Kotlin) projects, so I don't need to support an image where Java is included.
|
1.0
|
Find better ways to distribute the kobweb binary - ~~How to get onto package managers in Linux or Homebrew in Mac? What's the equivalent for Windows?~~
Support and/or intentionally reject the following package managers:
- [x] homebrew
- [x] scoop
- [X] ~~chocolatey~~ (1)
- [X] ~~gofish~~ (2)
- [X] ~~macports~~ (2)
- [X] ~~snapcraft~~ (3)
- [X] ~~spec~~ (4)
- [X] ~~docker~~ (5)
- [x] sdkman
* ~~strikethrough~~ means not now, we can close this bug without them, but I'd always be open to supporting them later
1. The process for Chocolatey feels a bit heavy handed for now. I can't simply self-host, instead I have to go through a whole process with their team, getting a review, etc. Plus, I found it hard to navigate the instructions. I'll leave this off for now, since Scoop is working, but can revisit this again in the future if Kobweb actually becomes regularly used.
2. Requires setting up PRs on specific repositories, not low hanging fruit
3. Strict mode is killing me. Snap sets up a whole environment just for Kobweb, making the download 100MB, and yet is getting killed by Snap and `snappy-debug` isn't really giving me useful information.
4. Not too familiar with Fedora or spec, but I'm considered adding support for it is diminishing returns and therefore isn't my highest priority.
5. A docker image is probably overkill. Users will be expected to have java on their machine anyway to compile Kobweb (Kotlin) projects, so I don't need to support an image where Java is included.
|
process
|
find better ways to distribute the kobweb binary how to get onto package managers in linux or homebrew in mac what s the equivalent for windows support and or intentionally reject the following package managers homebrew scoop chocolatey gofish macports snapcraft spec docker sdkman strikethrough means not now we can close this bug without them but i d always be open to supporting them later the process for chocolatey feels a bit heavy handed for now i can t simply self host instead i have to go through a whole process with their team getting a review etc plus i found it hard to navigate the instructions i ll leave this off for now since scoop is working but can revisit this again in the future if kobweb actually becomes regularly used requires setting up prs on specific repositories not low hanging fruit strict mode is killing me snap sets up a whole environment just for kobweb making the download and yet is getting killed by snap and snappy debug isn t really giving me useful information not too familiar with fedora or spec but i m considered adding support for it is diminishing returns and therefore isn t my highest priority a docker image is probably overkill users will be expected to have java on their machine anyway to compile kobweb kotlin projects so i don t need to support an image where java is included
| 1
|
203,604
| 15,885,578,535
|
IssuesEvent
|
2021-04-09 20:49:20
|
aws/aws-sdk-js-v3
|
https://api.github.com/repos/aws/aws-sdk-js-v3
|
closed
|
Roadmap / timeline to production release
|
documentation
|
**Describe the issue with documentation**
I couldn't see a timeline as to planned release candidates, maturity/features and target release dates.
**Additional context**
I'm trying to select between SDK versions. All I can see is that this SDK was announced in November 2018, but I've not seen a roadmap / timeline for other release candidates or final release.
|
1.0
|
Roadmap / timeline to production release - **Describe the issue with documentation**
I couldn't see a timeline as to planned release candidates, maturity/features and target release dates.
**Additional context**
I'm trying to select between SDK versions. All I can see is that this SDK was announced in November 2018, but I've not seen a roadmap / timeline for other release candidates or final release.
|
non_process
|
roadmap timeline to production release describe the issue with documentation i couldn t see a timeline as to planned release candidates maturity features and target release dates additional context i m trying to select between sdk versions all i can see is that this sdk was announced in november but i ve not seen a roadmap timeline for other release candidates or final release
| 0
|
265,146
| 28,244,697,717
|
IssuesEvent
|
2023-04-06 09:49:33
|
hshivhare67/platform_packages_apps_settings_AOSP10_r33
|
https://api.github.com/repos/hshivhare67/platform_packages_apps_settings_AOSP10_r33
|
closed
|
CVE-2022-20347 (High) detected in Settingsandroid-10.0.0_r44 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2022-20347 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r44</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/platform_packages_apps_settings_AOSP10_r33/commit/cdc44b74ac73d9c7eed82d7e753aba9efedac279">cdc44b74ac73d9c7eed82d7e753aba9efedac279</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/connecteddevice/ConnectedDeviceDashboardFragment.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In onAttach of ConnectedDeviceDashboardFragment.java, there is a possible permission bypass due to a confused deputy. This could lead to remote escalation of privilege in Bluetooth settings with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10 Android-11 Android-12 Android-12LAndroid ID: A-228450811
<p>Publish Date: 2022-08-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-20347>CVE-2022-20347</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-20347 (High) detected in Settingsandroid-10.0.0_r44 - autoclosed - ## CVE-2022-20347 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r44</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/platform_packages_apps_settings_AOSP10_r33/commit/cdc44b74ac73d9c7eed82d7e753aba9efedac279">cdc44b74ac73d9c7eed82d7e753aba9efedac279</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/connecteddevice/ConnectedDeviceDashboardFragment.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In onAttach of ConnectedDeviceDashboardFragment.java, there is a possible permission bypass due to a confused deputy. This could lead to remote escalation of privilege in Bluetooth settings with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10 Android-11 Android-12 Android-12LAndroid ID: A-228450811
<p>Publish Date: 2022-08-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-20347>CVE-2022-20347</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in settingsandroid autoclosed cve high severity vulnerability vulnerable library settingsandroid library home page a href found in head commit a href found in base branch main vulnerable source files src com android settings connecteddevice connecteddevicedashboardfragment java vulnerability details in onattach of connecteddevicedashboardfragment java there is a possible permission bypass due to a confused deputy this could lead to remote escalation of privilege in bluetooth settings with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with mend
| 0
|
20,791
| 27,534,682,491
|
IssuesEvent
|
2023-03-07 02:00:07
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Mon, 6 Mar 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### EcoTTA: Memory-Efficient Continual Test-time Adaptation via Self-distilled Regularization
- **Authors:** Junha Song, Jungsoo Lee, In So Kweon, Sungha Choi
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01904
- **Pdf link:** https://arxiv.org/pdf/2303.01904
- **Abstract**
This paper presents a simple yet effective approach that improves continual test-time adaptation (TTA) in a memory-efficient manner. TTA may primarily be conducted on edge devices with limited memory, so reducing memory is crucial but has been overlooked in previous TTA studies. In addition, long-term adaptation often leads to catastrophic forgetting and error accumulation, which hinders applying TTA in real-world deployments. Our approach consists of two components to address these issues. First, we present lightweight meta networks that can adapt the frozen original networks to the target domain. This novel architecture minimizes memory consumption by decreasing the size of intermediate activations required for backpropagation. Second, our novel self-distilled regularization controls the output of the meta networks not to deviate significantly from the output of the frozen original networks, thereby preserving well-trained knowledge from the source domain. Without additional memory, this regularization prevents error accumulation and catastrophic forgetting, resulting in stable performance even in long-term test-time adaptation. We demonstrate that our simple yet effective strategy outperforms other state-of-the-art methods on various benchmarks for image classification and semantic segmentation tasks. Notably, our proposed method with ResNet-50 and WideResNet-40 takes 86% and 80% less memory than the recent state-of-the-art method, CoTTA.
### Interruptions detection in video conferences
- **Authors:** Shmuel Horowitz, Dima Kagan, Galit Fuhrmann Alpert, Michael Fire
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Social and Information Networks (cs.SI)
- **Arxiv link:** https://arxiv.org/abs/2303.02052
- **Pdf link:** https://arxiv.org/pdf/2303.02052
- **Abstract**
In recent years, video conferencing (VC) popularity has skyrocketed for a wide range of activities. As a result, the number of VC users surged sharply. The sharp increase in VC usage has been accompanied by various newly emerging privacy and security challenges. VC meetings became a target for various security attacks, such as Zoombombing. Other VC-related challenges also emerged. For example, during COVID lockdowns, educators had to teach in online environments struggling with keeping students engaged for extended periods. In parallel, the amount of available VC videos has grown exponentially. Thus, users and companies are limited in finding abnormal segments in VC meetings within the converging volumes of data. Such abnormal events that affect most meeting participants may be indicators of interesting points in time, including security attacks or other changes in meeting climate, like someone joining a meeting or sharing a dramatic content. Here, we present a novel algorithm for detecting abnormal events in VC data. We curated VC publicly available recordings, including meetings with interruptions. We analyzed the videos using our algorithm, extracting time windows where abnormal occurrences were detected. Our algorithm is a pipeline that combines multiple methods in several steps to detect users' faces in each video frame, track face locations during the meeting and generate vector representations of a facial expression for each face in each frame. Vector representations are used to monitor changes in facial expressions throughout the meeting for each participant. The overall change in meeting climate is quantified using those parameters across all participants, and translating them into event anomaly detection. This is the first open pipeline for automatically detecting anomaly events in VC meetings. Our model detects abnormal events with 92.3% precision over the collected dataset.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### A Few-Shot Attention Recurrent Residual U-Net for Crack Segmentation
- **Authors:** Iason Katsamenis, Eftychios Protopapadakis, Nikolaos Bakalos, Anastasios Doulamis, Nikolaos Doulamis, Athanasios Voulodimos
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2303.01582
- **Pdf link:** https://arxiv.org/pdf/2303.01582
- **Abstract**
Recent studies indicate that deep learning plays a crucial role in the automated visual inspection of road infrastructures. However, current learning schemes are static, implying no dynamic adaptation to users' feedback. To address this drawback, we present a few-shot learning paradigm for the automated segmentation of road cracks, which is based on a U-Net architecture with recurrent residual and attention modules (R2AU-Net). The retraining strategy dynamically fine-tunes the weights of the U-Net as a few new rectified samples are being fed into the classifier. Extensive experiments show that the proposed few-shot R2AU-Net framework outperforms other state-of-the-art networks in terms of Dice and IoU metrics, on a new dataset, named CrackMap, which is made publicly available at https://github.com/ikatsamenis/CrackMap.
### Multi-Plane Neural Radiance Fields for Novel View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01736
- **Pdf link:** https://arxiv.org/pdf/2303.01736
- **Abstract**
Novel view synthesis is a long-standing problem that revolves around rendering frames of scenes from novel camera viewpoints. Volumetric approaches provide a solution for modeling occlusions through the explicit 3D representation of the camera frustum. Multi-plane Images (MPI) are volumetric methods that represent the scene using front-parallel planes at distinct depths but suffer from depth discretization leading to a 2.D scene representation. Another line of approach relies on implicit 3D scene representations. Neural Radiance Fields (NeRF) utilize neural networks for encapsulating the continuous 3D scene structure within the network weights achieving photorealistic synthesis results, however, methods are constrained to per-scene optimization settings which are inefficient in practice. Multi-plane Neural Radiance Fields (MINE) open the door for combining implicit and explicit scene representations. It enables continuous 3D scene representations, especially in the depth dimension, while utilizing the input image features to avoid per-scene optimization. The main drawback of the current literature work in this domain is being constrained to single-view input, limiting the synthesis ability to narrow viewpoint ranges. In this work, we thoroughly examine the performance, generalization, and efficiency of single-view multi-plane neural radiance fields. In addition, we propose a new multiplane NeRF architecture that accepts multiple views to improve the synthesis results and expand the viewing range. Features from the input source frames are effectively fused through a proposed attention-aware fusion module to highlight important information from different viewpoints. Experiments show the effectiveness of attention-based fusion and the promising outcomes of our proposed method when compared to multi-view NeRF and MPI techniques.
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Unsupervised 3D Shape Reconstruction by Part Retrieval and Assembly
- **Authors:** Xianghao Xu, Paul Guerrero, Matthew Fisher, Siddhartha Chaudhuri, Daniel Ritchie
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR)
- **Arxiv link:** https://arxiv.org/abs/2303.01999
- **Pdf link:** https://arxiv.org/pdf/2303.01999
- **Abstract**
Representing a 3D shape with a set of primitives can aid perception of structure, improve robotic object manipulation, and enable editing, stylization, and compression of 3D shapes. Existing methods either use simple parametric primitives or learn a generative shape space of parts. Both have limitations: parametric primitives lead to coarse approximations, while learned parts offer too little control over the decomposition. We instead propose to decompose shapes using a library of 3D parts provided by the user, giving full control over the choice of parts. The library can contain parts with high-quality geometry that are suitable for a given category, resulting in meaningful decompositions with clean geometry. The type of decomposition can also be controlled through the choice of parts in the library. Our method works via a self-supervised approach that iteratively retrieves parts from the library and refines their placements. We show that this approach gives higher reconstruction accuracy and more desirable decompositions than existing approaches. Additionally, we show how the decomposition can be controlled through the part library by using different part libraries to reconstruct the same shapes.
## Keyword: RAW
### A Few-Shot Attention Recurrent Residual U-Net for Crack Segmentation
- **Authors:** Iason Katsamenis, Eftychios Protopapadakis, Nikolaos Bakalos, Anastasios Doulamis, Nikolaos Doulamis, Athanasios Voulodimos
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2303.01582
- **Pdf link:** https://arxiv.org/pdf/2303.01582
- **Abstract**
Recent studies indicate that deep learning plays a crucial role in the automated visual inspection of road infrastructures. However, current learning schemes are static, implying no dynamic adaptation to users' feedback. To address this drawback, we present a few-shot learning paradigm for the automated segmentation of road cracks, which is based on a U-Net architecture with recurrent residual and attention modules (R2AU-Net). The retraining strategy dynamically fine-tunes the weights of the U-Net as a few new rectified samples are being fed into the classifier. Extensive experiments show that the proposed few-shot R2AU-Net framework outperforms other state-of-the-art networks in terms of Dice and IoU metrics, on a new dataset, named CrackMap, which is made publicly available at https://github.com/ikatsamenis/CrackMap.
### Towards Domain Generalization for Multi-view 3D Object Detection in Bird-Eye-View
- **Authors:** Shuo Wang, Xinhai Zhao, Hai-Ming Xu, Zehui Chen, Dameng Yu, Jiahao Chang, Zhen Yang, Feng Zhao
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01686
- **Pdf link:** https://arxiv.org/pdf/2303.01686
- **Abstract**
Multi-view 3D object detection (MV3D-Det) in Bird-Eye-View (BEV) has drawn extensive attention due to its low cost and high efficiency. Although new algorithms for camera-only 3D object detection have been continuously proposed, most of them may risk drastic performance degradation when the domain of input images differs from that of training. In this paper, we first analyze the causes of the domain gap for the MV3D-Det task. Based on the covariate shift assumption, we find that the gap mainly attributes to the feature distribution of BEV, which is determined by the quality of both depth estimation and 2D image's feature representation. To acquire a robust depth prediction, we propose to decouple the depth estimation from the intrinsic parameters of the camera (i.e. the focal length) through converting the prediction of metric depth to that of scale-invariant depth and perform dynamic perspective augmentation to increase the diversity of the extrinsic parameters (i.e. the camera poses) by utilizing homography. Moreover, we modify the focal length values to create multiple pseudo-domains and construct an adversarial training loss to encourage the feature representation to be more domain-agnostic. Without bells and whistles, our approach, namely DG-BEV, successfully alleviates the performance drop on the unseen target domain without impairing the accuracy of the source domain. Extensive experiments on various public datasets, including Waymo, nuScenes, and Lyft, demonstrate the generalization and effectiveness of our approach. To the best of our knowledge, this is the first systematic study to explore a domain generalization method for MV3D-Det.
### Multi-Plane Neural Radiance Fields for Novel View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01736
- **Pdf link:** https://arxiv.org/pdf/2303.01736
- **Abstract**
Novel view synthesis is a long-standing problem that revolves around rendering frames of scenes from novel camera viewpoints. Volumetric approaches provide a solution for modeling occlusions through the explicit 3D representation of the camera frustum. Multi-plane Images (MPI) are volumetric methods that represent the scene using front-parallel planes at distinct depths but suffer from depth discretization leading to a 2.D scene representation. Another line of approach relies on implicit 3D scene representations. Neural Radiance Fields (NeRF) utilize neural networks for encapsulating the continuous 3D scene structure within the network weights achieving photorealistic synthesis results, however, methods are constrained to per-scene optimization settings which are inefficient in practice. Multi-plane Neural Radiance Fields (MINE) open the door for combining implicit and explicit scene representations. It enables continuous 3D scene representations, especially in the depth dimension, while utilizing the input image features to avoid per-scene optimization. The main drawback of the current literature work in this domain is being constrained to single-view input, limiting the synthesis ability to narrow viewpoint ranges. In this work, we thoroughly examine the performance, generalization, and efficiency of single-view multi-plane neural radiance fields. In addition, we propose a new multiplane NeRF architecture that accepts multiple views to improve the synthesis results and expand the viewing range. Features from the input source frames are effectively fused through a proposed attention-aware fusion module to highlight important information from different viewpoints. Experiments show the effectiveness of attention-based fusion and the promising outcomes of our proposed method when compared to multi-view NeRF and MPI techniques.
### A Laplace-inspired Distribution on SO(3) for Probabilistic Rotation Estimation
- **Authors:** Yingda Yin, Yang Wang, He Wang, Baoquan Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01743
- **Pdf link:** https://arxiv.org/pdf/2303.01743
- **Abstract**
Estimating the 3DoF rotation from a single RGB image is an important yet challenging problem. Probabilistic rotation regression has raised more and more attention with the benefit of expressing uncertainty information along with the prediction. Though modeling noise using Gaussian-resembling Bingham distribution and matrix Fisher distribution is natural, they are shown to be sensitive to outliers for the nature of quadratic punishment to deviations. In this paper, we draw inspiration from multivariate Laplace distribution and propose a novel Rotation Laplace distribution on SO(3). Rotation Laplace distribution is robust to the disturbance of outliers and enforces much gradient to the low-error region, resulting in a better convergence. Our extensive experiments show that our proposed distribution achieves state-of-the-art performance for rotation regression tasks over both probabilistic and non-probabilistic baselines. Our project page is at https://pku-epic.github.io/RotationLaplace.
### Prior Information based Decomposition and Reconstruction Learning for Micro-Expression Recognition
- **Authors:** Jinsheng Wei, Haoyu Chen, Guanming Lu, Jingjie Yan, Yue Xie, Guoying Zhao
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01776
- **Pdf link:** https://arxiv.org/pdf/2303.01776
- **Abstract**
Micro-expression recognition (MER) draws intensive research interest as micro-expressions (MEs) can infer genuine emotions. Prior information can guide the model to learn discriminative ME features effectively. However, most works focus on researching the general models with a stronger representation ability to adaptively aggregate ME movement information in a holistic way, which may ignore the prior information and properties of MEs. To solve this issue, driven by the prior information that the category of ME can be inferred by the relationship between the actions of facial different components, this work designs a novel model that can conform to this prior information and learn ME movement features in an interpretable way. Specifically, this paper proposes a Decomposition and Reconstruction-based Graph Representation Learning (DeRe-GRL) model to effectively learn high-level ME features. DeRe-GRL includes two modules: Action Decomposition Module (ADM) and Relation Reconstruction Module (RRM), where ADM learns action features of facial key components and RRM explores the relationship between these action features. Based on facial key components, ADM divides the geometric movement features extracted by the graph model-based backbone into several sub-features, and learns the map matrix to map these sub-features into multiple action features; then, RRM learns weights to weight all action features to build the relationship between action features. The experimental results demonstrate the effectiveness of the proposed modules, and the proposed method achieves competitive performance.
### BSH-Det3D: Improving 3D Object Detection with BEV Shape Heatmap
- **Authors:** You Shen, Yunzhou Zhang, Yanmin Wu, Zhenyu Wang, Linghao Yang, Sonya Coleman, Dermot Kerr
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.02000
- **Pdf link:** https://arxiv.org/pdf/2303.02000
- **Abstract**
The progress of LiDAR-based 3D object detection has significantly enhanced developments in autonomous driving and robotics. However, due to the limitations of LiDAR sensors, object shapes suffer from deterioration in occluded and distant areas, which creates a fundamental challenge to 3D perception. Existing methods estimate specific 3D shapes and achieve remarkable performance. However, these methods rely on extensive computation and memory, causing imbalances between accuracy and real-time performance. To tackle this challenge, we propose a novel LiDAR-based 3D object detection model named BSH-Det3D, which applies an effective way to enhance spatial features by estimating complete shapes from a bird's eye view (BEV). Specifically, we design the Pillar-based Shape Completion (PSC) module to predict the probability of occupancy whether a pillar contains object shapes. The PSC module generates a BEV shape heatmap for each scene. After integrating with heatmaps, BSH-Det3D can provide additional information in shape deterioration areas and generate high-quality 3D proposals. We also design an attention-based densification fusion module (ADF) to adaptively associate the sparse features with heatmaps and raw points. The ADF module integrates the advantages of points and shapes knowledge with negligible overheads. Extensive experiments on the KITTI benchmark achieve state-of-the-art (SOTA) performance in terms of accuracy and speed, demonstrating the efficiency and flexibility of BSH-Det3D. The source code is available on https://github.com/mystorm16/BSH-Det3D.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Mon, 6 Mar 23 - ## Keyword: events
### EcoTTA: Memory-Efficient Continual Test-time Adaptation via Self-distilled Regularization
- **Authors:** Junha Song, Jungsoo Lee, In So Kweon, Sungha Choi
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01904
- **Pdf link:** https://arxiv.org/pdf/2303.01904
- **Abstract**
This paper presents a simple yet effective approach that improves continual test-time adaptation (TTA) in a memory-efficient manner. TTA may primarily be conducted on edge devices with limited memory, so reducing memory is crucial but has been overlooked in previous TTA studies. In addition, long-term adaptation often leads to catastrophic forgetting and error accumulation, which hinders applying TTA in real-world deployments. Our approach consists of two components to address these issues. First, we present lightweight meta networks that can adapt the frozen original networks to the target domain. This novel architecture minimizes memory consumption by decreasing the size of intermediate activations required for backpropagation. Second, our novel self-distilled regularization controls the output of the meta networks not to deviate significantly from the output of the frozen original networks, thereby preserving well-trained knowledge from the source domain. Without additional memory, this regularization prevents error accumulation and catastrophic forgetting, resulting in stable performance even in long-term test-time adaptation. We demonstrate that our simple yet effective strategy outperforms other state-of-the-art methods on various benchmarks for image classification and semantic segmentation tasks. Notably, our proposed method with ResNet-50 and WideResNet-40 takes 86% and 80% less memory than the recent state-of-the-art method, CoTTA.
### Interruptions detection in video conferences
- **Authors:** Shmuel Horowitz, Dima Kagan, Galit Fuhrmann Alpert, Michael Fire
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Social and Information Networks (cs.SI)
- **Arxiv link:** https://arxiv.org/abs/2303.02052
- **Pdf link:** https://arxiv.org/pdf/2303.02052
- **Abstract**
In recent years, video conferencing (VC) popularity has skyrocketed for a wide range of activities. As a result, the number of VC users surged sharply. The sharp increase in VC usage has been accompanied by various newly emerging privacy and security challenges. VC meetings became a target for various security attacks, such as Zoombombing. Other VC-related challenges also emerged. For example, during COVID lockdowns, educators had to teach in online environments struggling with keeping students engaged for extended periods. In parallel, the amount of available VC videos has grown exponentially. Thus, users and companies are limited in finding abnormal segments in VC meetings within the converging volumes of data. Such abnormal events that affect most meeting participants may be indicators of interesting points in time, including security attacks or other changes in meeting climate, like someone joining a meeting or sharing a dramatic content. Here, we present a novel algorithm for detecting abnormal events in VC data. We curated VC publicly available recordings, including meetings with interruptions. We analyzed the videos using our algorithm, extracting time windows where abnormal occurrences were detected. Our algorithm is a pipeline that combines multiple methods in several steps to detect users' faces in each video frame, track face locations during the meeting and generate vector representations of a facial expression for each face in each frame. Vector representations are used to monitor changes in facial expressions throughout the meeting for each participant. The overall change in meeting climate is quantified using those parameters across all participants, and translating them into event anomaly detection. This is the first open pipeline for automatically detecting anomaly events in VC meetings. Our model detects abnormal events with 92.3% precision over the collected dataset.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### A Few-Shot Attention Recurrent Residual U-Net for Crack Segmentation
- **Authors:** Iason Katsamenis, Eftychios Protopapadakis, Nikolaos Bakalos, Anastasios Doulamis, Nikolaos Doulamis, Athanasios Voulodimos
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2303.01582
- **Pdf link:** https://arxiv.org/pdf/2303.01582
- **Abstract**
Recent studies indicate that deep learning plays a crucial role in the automated visual inspection of road infrastructures. However, current learning schemes are static, implying no dynamic adaptation to users' feedback. To address this drawback, we present a few-shot learning paradigm for the automated segmentation of road cracks, which is based on a U-Net architecture with recurrent residual and attention modules (R2AU-Net). The retraining strategy dynamically fine-tunes the weights of the U-Net as a few new rectified samples are being fed into the classifier. Extensive experiments show that the proposed few-shot R2AU-Net framework outperforms other state-of-the-art networks in terms of Dice and IoU metrics, on a new dataset, named CrackMap, which is made publicly available at https://github.com/ikatsamenis/CrackMap.
### Multi-Plane Neural Radiance Fields for Novel View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01736
- **Pdf link:** https://arxiv.org/pdf/2303.01736
- **Abstract**
Novel view synthesis is a long-standing problem that revolves around rendering frames of scenes from novel camera viewpoints. Volumetric approaches provide a solution for modeling occlusions through the explicit 3D representation of the camera frustum. Multi-plane Images (MPI) are volumetric methods that represent the scene using front-parallel planes at distinct depths but suffer from depth discretization leading to a 2.D scene representation. Another line of approach relies on implicit 3D scene representations. Neural Radiance Fields (NeRF) utilize neural networks for encapsulating the continuous 3D scene structure within the network weights achieving photorealistic synthesis results, however, methods are constrained to per-scene optimization settings which are inefficient in practice. Multi-plane Neural Radiance Fields (MINE) open the door for combining implicit and explicit scene representations. It enables continuous 3D scene representations, especially in the depth dimension, while utilizing the input image features to avoid per-scene optimization. The main drawback of the current literature work in this domain is being constrained to single-view input, limiting the synthesis ability to narrow viewpoint ranges. In this work, we thoroughly examine the performance, generalization, and efficiency of single-view multi-plane neural radiance fields. In addition, we propose a new multiplane NeRF architecture that accepts multiple views to improve the synthesis results and expand the viewing range. Features from the input source frames are effectively fused through a proposed attention-aware fusion module to highlight important information from different viewpoints. Experiments show the effectiveness of attention-based fusion and the promising outcomes of our proposed method when compared to multi-view NeRF and MPI techniques.
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Unsupervised 3D Shape Reconstruction by Part Retrieval and Assembly
- **Authors:** Xianghao Xu, Paul Guerrero, Matthew Fisher, Siddhartha Chaudhuri, Daniel Ritchie
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR)
- **Arxiv link:** https://arxiv.org/abs/2303.01999
- **Pdf link:** https://arxiv.org/pdf/2303.01999
- **Abstract**
Representing a 3D shape with a set of primitives can aid perception of structure, improve robotic object manipulation, and enable editing, stylization, and compression of 3D shapes. Existing methods either use simple parametric primitives or learn a generative shape space of parts. Both have limitations: parametric primitives lead to coarse approximations, while learned parts offer too little control over the decomposition. We instead propose to decompose shapes using a library of 3D parts provided by the user, giving full control over the choice of parts. The library can contain parts with high-quality geometry that are suitable for a given category, resulting in meaningful decompositions with clean geometry. The type of decomposition can also be controlled through the choice of parts in the library. Our method works via a self-supervised approach that iteratively retrieves parts from the library and refines their placements. We show that this approach gives higher reconstruction accuracy and more desirable decompositions than existing approaches. Additionally, we show how the decomposition can be controlled through the part library by using different part libraries to reconstruct the same shapes.
## Keyword: RAW
### A Few-Shot Attention Recurrent Residual U-Net for Crack Segmentation
- **Authors:** Iason Katsamenis, Eftychios Protopapadakis, Nikolaos Bakalos, Anastasios Doulamis, Nikolaos Doulamis, Athanasios Voulodimos
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2303.01582
- **Pdf link:** https://arxiv.org/pdf/2303.01582
- **Abstract**
Recent studies indicate that deep learning plays a crucial role in the automated visual inspection of road infrastructures. However, current learning schemes are static, implying no dynamic adaptation to users' feedback. To address this drawback, we present a few-shot learning paradigm for the automated segmentation of road cracks, which is based on a U-Net architecture with recurrent residual and attention modules (R2AU-Net). The retraining strategy dynamically fine-tunes the weights of the U-Net as a few new rectified samples are being fed into the classifier. Extensive experiments show that the proposed few-shot R2AU-Net framework outperforms other state-of-the-art networks in terms of Dice and IoU metrics, on a new dataset, named CrackMap, which is made publicly available at https://github.com/ikatsamenis/CrackMap.
### Towards Domain Generalization for Multi-view 3D Object Detection in Bird-Eye-View
- **Authors:** Shuo Wang, Xinhai Zhao, Hai-Ming Xu, Zehui Chen, Dameng Yu, Jiahao Chang, Zhen Yang, Feng Zhao
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01686
- **Pdf link:** https://arxiv.org/pdf/2303.01686
- **Abstract**
Multi-view 3D object detection (MV3D-Det) in Bird-Eye-View (BEV) has drawn extensive attention due to its low cost and high efficiency. Although new algorithms for camera-only 3D object detection have been continuously proposed, most of them may risk drastic performance degradation when the domain of input images differs from that of training. In this paper, we first analyze the causes of the domain gap for the MV3D-Det task. Based on the covariate shift assumption, we find that the gap mainly attributes to the feature distribution of BEV, which is determined by the quality of both depth estimation and 2D image's feature representation. To acquire a robust depth prediction, we propose to decouple the depth estimation from the intrinsic parameters of the camera (i.e. the focal length) through converting the prediction of metric depth to that of scale-invariant depth and perform dynamic perspective augmentation to increase the diversity of the extrinsic parameters (i.e. the camera poses) by utilizing homography. Moreover, we modify the focal length values to create multiple pseudo-domains and construct an adversarial training loss to encourage the feature representation to be more domain-agnostic. Without bells and whistles, our approach, namely DG-BEV, successfully alleviates the performance drop on the unseen target domain without impairing the accuracy of the source domain. Extensive experiments on various public datasets, including Waymo, nuScenes, and Lyft, demonstrate the generalization and effectiveness of our approach. To the best of our knowledge, this is the first systematic study to explore a domain generalization method for MV3D-Det.
### Multi-Plane Neural Radiance Fields for Novel View Synthesis
- **Authors:** Youssef Abdelkareem, Shady Shehata, Fakhri Karray
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01736
- **Pdf link:** https://arxiv.org/pdf/2303.01736
- **Abstract**
Novel view synthesis is a long-standing problem that revolves around rendering frames of scenes from novel camera viewpoints. Volumetric approaches provide a solution for modeling occlusions through the explicit 3D representation of the camera frustum. Multi-plane Images (MPI) are volumetric methods that represent the scene using front-parallel planes at distinct depths but suffer from depth discretization leading to a 2.D scene representation. Another line of approach relies on implicit 3D scene representations. Neural Radiance Fields (NeRF) utilize neural networks for encapsulating the continuous 3D scene structure within the network weights achieving photorealistic synthesis results, however, methods are constrained to per-scene optimization settings which are inefficient in practice. Multi-plane Neural Radiance Fields (MINE) open the door for combining implicit and explicit scene representations. It enables continuous 3D scene representations, especially in the depth dimension, while utilizing the input image features to avoid per-scene optimization. The main drawback of the current literature work in this domain is being constrained to single-view input, limiting the synthesis ability to narrow viewpoint ranges. In this work, we thoroughly examine the performance, generalization, and efficiency of single-view multi-plane neural radiance fields. In addition, we propose a new multiplane NeRF architecture that accepts multiple views to improve the synthesis results and expand the viewing range. Features from the input source frames are effectively fused through a proposed attention-aware fusion module to highlight important information from different viewpoints. Experiments show the effectiveness of attention-based fusion and the promising outcomes of our proposed method when compared to multi-view NeRF and MPI techniques.
### A Laplace-inspired Distribution on SO(3) for Probabilistic Rotation Estimation
- **Authors:** Yingda Yin, Yang Wang, He Wang, Baoquan Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01743
- **Pdf link:** https://arxiv.org/pdf/2303.01743
- **Abstract**
Estimating the 3DoF rotation from a single RGB image is an important yet challenging problem. Probabilistic rotation regression has raised more and more attention with the benefit of expressing uncertainty information along with the prediction. Though modeling noise using Gaussian-resembling Bingham distribution and matrix Fisher distribution is natural, they are shown to be sensitive to outliers for the nature of quadratic punishment to deviations. In this paper, we draw inspiration from multivariate Laplace distribution and propose a novel Rotation Laplace distribution on SO(3). Rotation Laplace distribution is robust to the disturbance of outliers and enforces much gradient to the low-error region, resulting in a better convergence. Our extensive experiments show that our proposed distribution achieves state-of-the-art performance for rotation regression tasks over both probabilistic and non-probabilistic baselines. Our project page is at https://pku-epic.github.io/RotationLaplace.
### Prior Information based Decomposition and Reconstruction Learning for Micro-Expression Recognition
- **Authors:** Jinsheng Wei, Haoyu Chen, Guanming Lu, Jingjie Yan, Yue Xie, Guoying Zhao
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.01776
- **Pdf link:** https://arxiv.org/pdf/2303.01776
- **Abstract**
Micro-expression recognition (MER) draws intensive research interest as micro-expressions (MEs) can infer genuine emotions. Prior information can guide the model to learn discriminative ME features effectively. However, most works focus on researching the general models with a stronger representation ability to adaptively aggregate ME movement information in a holistic way, which may ignore the prior information and properties of MEs. To solve this issue, driven by the prior information that the category of ME can be inferred by the relationship between the actions of facial different components, this work designs a novel model that can conform to this prior information and learn ME movement features in an interpretable way. Specifically, this paper proposes a Decomposition and Reconstruction-based Graph Representation Learning (DeRe-GRL) model to effectively learn high-level ME features. DeRe-GRL includes two modules: Action Decomposition Module (ADM) and Relation Reconstruction Module (RRM), where ADM learns action features of facial key components and RRM explores the relationship between these action features. Based on facial key components, ADM divides the geometric movement features extracted by the graph model-based backbone into several sub-features, and learns the map matrix to map these sub-features into multiple action features; then, RRM learns weights to weight all action features to build the relationship between action features. The experimental results demonstrate the effectiveness of the proposed modules, and the proposed method achieves competitive performance.
### BSH-Det3D: Improving 3D Object Detection with BEV Shape Heatmap
- **Authors:** You Shen, Yunzhou Zhang, Yanmin Wu, Zhenyu Wang, Linghao Yang, Sonya Coleman, Dermot Kerr
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2303.02000
- **Pdf link:** https://arxiv.org/pdf/2303.02000
- **Abstract**
The progress of LiDAR-based 3D object detection has significantly enhanced developments in autonomous driving and robotics. However, due to the limitations of LiDAR sensors, object shapes suffer from deterioration in occluded and distant areas, which creates a fundamental challenge to 3D perception. Existing methods estimate specific 3D shapes and achieve remarkable performance. However, these methods rely on extensive computation and memory, causing imbalances between accuracy and real-time performance. To tackle this challenge, we propose a novel LiDAR-based 3D object detection model named BSH-Det3D, which applies an effective way to enhance spatial features by estimating complete shapes from a bird's eye view (BEV). Specifically, we design the Pillar-based Shape Completion (PSC) module to predict the probability of occupancy whether a pillar contains object shapes. The PSC module generates a BEV shape heatmap for each scene. After integrating with heatmaps, BSH-Det3D can provide additional information in shape deterioration areas and generate high-quality 3D proposals. We also design an attention-based densification fusion module (ADF) to adaptively associate the sparse features with heatmaps and raw points. The ADF module integrates the advantages of points and shapes knowledge with negligible overheads. Extensive experiments on the KITTI benchmark achieve state-of-the-art (SOTA) performance in terms of accuracy and speed, demonstrating the efficiency and flexibility of BSH-Det3D. The source code is available on https://github.com/mystorm16/BSH-Det3D.
## Keyword: raw image
There is no result
|
process
|
new submissions for mon mar keyword events ecotta memory efficient continual test time adaptation via self distilled regularization authors junha song jungsoo lee in so kweon sungha choi subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract this paper presents a simple yet effective approach that improves continual test time adaptation tta in a memory efficient manner tta may primarily be conducted on edge devices with limited memory so reducing memory is crucial but has been overlooked in previous tta studies in addition long term adaptation often leads to catastrophic forgetting and error accumulation which hinders applying tta in real world deployments our approach consists of two components to address these issues first we present lightweight meta networks that can adapt the frozen original networks to the target domain this novel architecture minimizes memory consumption by decreasing the size of intermediate activations required for backpropagation second our novel self distilled regularization controls the output of the meta networks not to deviate significantly from the output of the frozen original networks thereby preserving well trained knowledge from the source domain without additional memory this regularization prevents error accumulation and catastrophic forgetting resulting in stable performance even in long term test time adaptation we demonstrate that our simple yet effective strategy outperforms other state of the art methods on various benchmarks for image classification and semantic segmentation tasks notably our proposed method with resnet and wideresnet takes and less memory than the recent state of the art method cotta interruptions detection in video conferences authors shmuel horowitz dima kagan galit fuhrmann alpert michael fire subjects computer vision and pattern recognition cs cv social and information networks cs si arxiv link pdf link abstract in recent years video conferencing vc popularity has skyrocketed for a wide range of activities as a result the number of vc users surged sharply the sharp increase in vc usage has been accompanied by various newly emerging privacy and security challenges vc meetings became a target for various security attacks such as zoombombing other vc related challenges also emerged for example during covid lockdowns educators had to teach in online environments struggling with keeping students engaged for extended periods in parallel the amount of available vc videos has grown exponentially thus users and companies are limited in finding abnormal segments in vc meetings within the converging volumes of data such abnormal events that affect most meeting participants may be indicators of interesting points in time including security attacks or other changes in meeting climate like someone joining a meeting or sharing a dramatic content here we present a novel algorithm for detecting abnormal events in vc data we curated vc publicly available recordings including meetings with interruptions we analyzed the videos using our algorithm extracting time windows where abnormal occurrences were detected our algorithm is a pipeline that combines multiple methods in several steps to detect users faces in each video frame track face locations during the meeting and generate vector representations of a facial expression for each face in each frame vector representations are used to monitor changes in facial expressions throughout the meeting for each participant the overall change in meeting climate is quantified using those parameters across all participants and translating them into event anomaly detection this is the first open pipeline for automatically detecting anomaly events in vc meetings our model detects abnormal events with precision over the collected dataset keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb a few shot attention recurrent residual u net for crack segmentation authors iason katsamenis eftychios protopapadakis nikolaos bakalos anastasios doulamis nikolaos doulamis athanasios voulodimos subjects computer vision and pattern recognition cs cv machine learning cs lg image and video processing eess iv arxiv link pdf link abstract recent studies indicate that deep learning plays a crucial role in the automated visual inspection of road infrastructures however current learning schemes are static implying no dynamic adaptation to users feedback to address this drawback we present a few shot learning paradigm for the automated segmentation of road cracks which is based on a u net architecture with recurrent residual and attention modules net the retraining strategy dynamically fine tunes the weights of the u net as a few new rectified samples are being fed into the classifier extensive experiments show that the proposed few shot net framework outperforms other state of the art networks in terms of dice and iou metrics on a new dataset named crackmap which is made publicly available at multi plane neural radiance fields for novel view synthesis authors youssef abdelkareem shady shehata fakhri karray subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract novel view synthesis is a long standing problem that revolves around rendering frames of scenes from novel camera viewpoints volumetric approaches provide a solution for modeling occlusions through the explicit representation of the camera frustum multi plane images mpi are volumetric methods that represent the scene using front parallel planes at distinct depths but suffer from depth discretization leading to a d scene representation another line of approach relies on implicit scene representations neural radiance fields nerf utilize neural networks for encapsulating the continuous scene structure within the network weights achieving photorealistic synthesis results however methods are constrained to per scene optimization settings which are inefficient in practice multi plane neural radiance fields mine open the door for combining implicit and explicit scene representations it enables continuous scene representations especially in the depth dimension while utilizing the input image features to avoid per scene optimization the main drawback of the current literature work in this domain is being constrained to single view input limiting the synthesis ability to narrow viewpoint ranges in this work we thoroughly examine the performance generalization and efficiency of single view multi plane neural radiance fields in addition we propose a new multiplane nerf architecture that accepts multiple views to improve the synthesis results and expand the viewing range features from the input source frames are effectively fused through a proposed attention aware fusion module to highlight important information from different viewpoints experiments show the effectiveness of attention based fusion and the promising outcomes of our proposed method when compared to multi view nerf and mpi techniques keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression unsupervised shape reconstruction by part retrieval and assembly authors xianghao xu paul guerrero matthew fisher siddhartha chaudhuri daniel ritchie subjects computer vision and pattern recognition cs cv graphics cs gr arxiv link pdf link abstract representing a shape with a set of primitives can aid perception of structure improve robotic object manipulation and enable editing stylization and compression of shapes existing methods either use simple parametric primitives or learn a generative shape space of parts both have limitations parametric primitives lead to coarse approximations while learned parts offer too little control over the decomposition we instead propose to decompose shapes using a library of parts provided by the user giving full control over the choice of parts the library can contain parts with high quality geometry that are suitable for a given category resulting in meaningful decompositions with clean geometry the type of decomposition can also be controlled through the choice of parts in the library our method works via a self supervised approach that iteratively retrieves parts from the library and refines their placements we show that this approach gives higher reconstruction accuracy and more desirable decompositions than existing approaches additionally we show how the decomposition can be controlled through the part library by using different part libraries to reconstruct the same shapes keyword raw a few shot attention recurrent residual u net for crack segmentation authors iason katsamenis eftychios protopapadakis nikolaos bakalos anastasios doulamis nikolaos doulamis athanasios voulodimos subjects computer vision and pattern recognition cs cv machine learning cs lg image and video processing eess iv arxiv link pdf link abstract recent studies indicate that deep learning plays a crucial role in the automated visual inspection of road infrastructures however current learning schemes are static implying no dynamic adaptation to users feedback to address this drawback we present a few shot learning paradigm for the automated segmentation of road cracks which is based on a u net architecture with recurrent residual and attention modules net the retraining strategy dynamically fine tunes the weights of the u net as a few new rectified samples are being fed into the classifier extensive experiments show that the proposed few shot net framework outperforms other state of the art networks in terms of dice and iou metrics on a new dataset named crackmap which is made publicly available at towards domain generalization for multi view object detection in bird eye view authors shuo wang xinhai zhao hai ming xu zehui chen dameng yu jiahao chang zhen yang feng zhao subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract multi view object detection det in bird eye view bev has drawn extensive attention due to its low cost and high efficiency although new algorithms for camera only object detection have been continuously proposed most of them may risk drastic performance degradation when the domain of input images differs from that of training in this paper we first analyze the causes of the domain gap for the det task based on the covariate shift assumption we find that the gap mainly attributes to the feature distribution of bev which is determined by the quality of both depth estimation and image s feature representation to acquire a robust depth prediction we propose to decouple the depth estimation from the intrinsic parameters of the camera i e the focal length through converting the prediction of metric depth to that of scale invariant depth and perform dynamic perspective augmentation to increase the diversity of the extrinsic parameters i e the camera poses by utilizing homography moreover we modify the focal length values to create multiple pseudo domains and construct an adversarial training loss to encourage the feature representation to be more domain agnostic without bells and whistles our approach namely dg bev successfully alleviates the performance drop on the unseen target domain without impairing the accuracy of the source domain extensive experiments on various public datasets including waymo nuscenes and lyft demonstrate the generalization and effectiveness of our approach to the best of our knowledge this is the first systematic study to explore a domain generalization method for det multi plane neural radiance fields for novel view synthesis authors youssef abdelkareem shady shehata fakhri karray subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract novel view synthesis is a long standing problem that revolves around rendering frames of scenes from novel camera viewpoints volumetric approaches provide a solution for modeling occlusions through the explicit representation of the camera frustum multi plane images mpi are volumetric methods that represent the scene using front parallel planes at distinct depths but suffer from depth discretization leading to a d scene representation another line of approach relies on implicit scene representations neural radiance fields nerf utilize neural networks for encapsulating the continuous scene structure within the network weights achieving photorealistic synthesis results however methods are constrained to per scene optimization settings which are inefficient in practice multi plane neural radiance fields mine open the door for combining implicit and explicit scene representations it enables continuous scene representations especially in the depth dimension while utilizing the input image features to avoid per scene optimization the main drawback of the current literature work in this domain is being constrained to single view input limiting the synthesis ability to narrow viewpoint ranges in this work we thoroughly examine the performance generalization and efficiency of single view multi plane neural radiance fields in addition we propose a new multiplane nerf architecture that accepts multiple views to improve the synthesis results and expand the viewing range features from the input source frames are effectively fused through a proposed attention aware fusion module to highlight important information from different viewpoints experiments show the effectiveness of attention based fusion and the promising outcomes of our proposed method when compared to multi view nerf and mpi techniques a laplace inspired distribution on so for probabilistic rotation estimation authors yingda yin yang wang he wang baoquan chen subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract estimating the rotation from a single rgb image is an important yet challenging problem probabilistic rotation regression has raised more and more attention with the benefit of expressing uncertainty information along with the prediction though modeling noise using gaussian resembling bingham distribution and matrix fisher distribution is natural they are shown to be sensitive to outliers for the nature of quadratic punishment to deviations in this paper we draw inspiration from multivariate laplace distribution and propose a novel rotation laplace distribution on so rotation laplace distribution is robust to the disturbance of outliers and enforces much gradient to the low error region resulting in a better convergence our extensive experiments show that our proposed distribution achieves state of the art performance for rotation regression tasks over both probabilistic and non probabilistic baselines our project page is at prior information based decomposition and reconstruction learning for micro expression recognition authors jinsheng wei haoyu chen guanming lu jingjie yan yue xie guoying zhao subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract micro expression recognition mer draws intensive research interest as micro expressions mes can infer genuine emotions prior information can guide the model to learn discriminative me features effectively however most works focus on researching the general models with a stronger representation ability to adaptively aggregate me movement information in a holistic way which may ignore the prior information and properties of mes to solve this issue driven by the prior information that the category of me can be inferred by the relationship between the actions of facial different components this work designs a novel model that can conform to this prior information and learn me movement features in an interpretable way specifically this paper proposes a decomposition and reconstruction based graph representation learning dere grl model to effectively learn high level me features dere grl includes two modules action decomposition module adm and relation reconstruction module rrm where adm learns action features of facial key components and rrm explores the relationship between these action features based on facial key components adm divides the geometric movement features extracted by the graph model based backbone into several sub features and learns the map matrix to map these sub features into multiple action features then rrm learns weights to weight all action features to build the relationship between action features the experimental results demonstrate the effectiveness of the proposed modules and the proposed method achieves competitive performance bsh improving object detection with bev shape heatmap authors you shen yunzhou zhang yanmin wu zhenyu wang linghao yang sonya coleman dermot kerr subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the progress of lidar based object detection has significantly enhanced developments in autonomous driving and robotics however due to the limitations of lidar sensors object shapes suffer from deterioration in occluded and distant areas which creates a fundamental challenge to perception existing methods estimate specific shapes and achieve remarkable performance however these methods rely on extensive computation and memory causing imbalances between accuracy and real time performance to tackle this challenge we propose a novel lidar based object detection model named bsh which applies an effective way to enhance spatial features by estimating complete shapes from a bird s eye view bev specifically we design the pillar based shape completion psc module to predict the probability of occupancy whether a pillar contains object shapes the psc module generates a bev shape heatmap for each scene after integrating with heatmaps bsh can provide additional information in shape deterioration areas and generate high quality proposals we also design an attention based densification fusion module adf to adaptively associate the sparse features with heatmaps and raw points the adf module integrates the advantages of points and shapes knowledge with negligible overheads extensive experiments on the kitti benchmark achieve state of the art sota performance in terms of accuracy and speed demonstrating the efficiency and flexibility of bsh the source code is available on keyword raw image there is no result
| 1
|
63,862
| 18,022,024,876
|
IssuesEvent
|
2021-09-16 20:47:25
|
department-of-veterans-affairs/vets-design-system-documentation
|
https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation
|
closed
|
[TESTING]: OMB Info - Unit tests MUST run axe checks with the modal open for better code coverage
|
pattern-update 508-defect-3 accessibility testing vsp-design-system-team
|
@1Copenut commented on [Wed Jun 03 2020](https://github.com/department-of-veterans-affairs/va.gov-team/issues/9816)
## [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
**Feedback framework**
- **βοΈ Must** for if the feedback must be applied
- **β οΈ Should** if the feedback is best practice
- **βοΈ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
The OMB Info component includes a modal window of legal text. The unit tests don't run an axe check on this modal. The component depends on this modal, so it must include an axe check with the modal open. This looks like it can be passed as a prop easily in the `OMBInfo.unit.spec.jsx` file.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] Unit test file includes an axe check with the modal window open
- [ ] Axe checks return 0 violations
## Environment
* https://department-of-veterans-affairs.github.io/veteran-facing-services-tools/visual-design/components/ombinfo/
|
1.0
|
[TESTING]: OMB Info - Unit tests MUST run axe checks with the modal open for better code coverage - @1Copenut commented on [Wed Jun 03 2020](https://github.com/department-of-veterans-affairs/va.gov-team/issues/9816)
## [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
**Feedback framework**
- **βοΈ Must** for if the feedback must be applied
- **β οΈ Should** if the feedback is best practice
- **βοΈ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
The OMB Info component includes a modal window of legal text. The unit tests don't run an axe check on this modal. The component depends on this modal, so it must include an axe check with the modal open. This looks like it can be passed as a prop easily in the `OMBInfo.unit.spec.jsx` file.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] Unit test file includes an axe check with the modal window open
- [ ] Axe checks return 0 violations
## Environment
* https://department-of-veterans-affairs.github.io/veteran-facing-services-tools/visual-design/components/ombinfo/
|
non_process
|
omb info unit tests must run axe checks with the modal open for better code coverage commented on enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework βοΈ must for if the feedback must be applied β οΈ should if the feedback is best practice βοΈ consider for suggestions enhancements description the omb info component includes a modal window of legal text the unit tests don t run an axe check on this modal the component depends on this modal so it must include an axe check with the modal open this looks like it can be passed as a prop easily in the ombinfo unit spec jsx file point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact trevor acceptance criteria unit test file includes an axe check with the modal window open axe checks return violations environment
| 0
|
413,622
| 12,076,666,918
|
IssuesEvent
|
2020-04-17 08:04:02
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
All menus and tables are magnified :/
|
Bug Feedback High Priority
|
Author Name: **Shiva Raissi** (Shiva Raissi)
Original Redmine Issue: [22114](https://issues.qgis.org/issues/22114)
Affected QGIS version: 3.6.2
Redmine category:unknown
---
Hi,
Since yesterday, all menus and tables are magnified in QGIS. I tried uninstalling and reinstalling the application but that didn't solve the issue. I'm using the 3.6.2 version and didn't have this issue before updating to this version.
I would appreciate if someone can help me with this.
Thanks.
---
- [QGIS Bug.PNG](https://issues.qgis.org/attachments/download/14968/QGIS%20Bug.PNG) (Shiva Raissi)
|
1.0
|
All menus and tables are magnified :/ - Author Name: **Shiva Raissi** (Shiva Raissi)
Original Redmine Issue: [22114](https://issues.qgis.org/issues/22114)
Affected QGIS version: 3.6.2
Redmine category:unknown
---
Hi,
Since yesterday, all menus and tables are magnified in QGIS. I tried uninstalling and reinstalling the application but that didn't solve the issue. I'm using the 3.6.2 version and didn't have this issue before updating to this version.
I would appreciate if someone can help me with this.
Thanks.
---
- [QGIS Bug.PNG](https://issues.qgis.org/attachments/download/14968/QGIS%20Bug.PNG) (Shiva Raissi)
|
non_process
|
all menus and tables are magnified author name shiva raissi shiva raissi original redmine issue affected qgis version redmine category unknown hi since yesterday all menus and tables are magnified in qgis i tried uninstalling and reinstalling the application but that didn t solve the issue i m using the version and didn t have this issue before updating to this version i would appreciate if someone can help me with this thanks shiva raissi
| 0
|
22,101
| 30,631,062,097
|
IssuesEvent
|
2023-07-24 14:34:15
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
PATH mutation using EnvironmentVariableCollection prepend is overwritten in zsh
|
terminal-process
|
- VS Code Version: 1.80.1
- OS Version: Darwin x64 22.5.0
VS Code Go extension tries to change the integrated terminals' PATH environment variable using `EnvironmentVariableCollection.prepend` api.
I verified that the change contribution is known to vscode using "Terminal: Show Environment Contributions" command.
```
# Terminal Environment Changes
## Extension: golang.go
- `PATH=/Users/hakim/sdk/go1.20.3/bin:${env:PATH}`
## Extension: vscode.git
- `GIT_ASKPASS=/Applications/Visual Studio Code.app/Contents/Resources/app/extensions/git/dist/askpass.sh`
...
```
When I create a terminal tab with `bash`, I see the PATH change applied and go sdk path prepended as expected.
But, when I create a terminal with `zsh`, it looks like zsh prepends the login shell PATH again after applying our extension's PATH change.
Note all `"terminal.integrated.profiles.*"` settings were default.
A workaround I found is to remove `"-l"` arg from the default `"terminal.integrated.profiles.osx.zsh"` setting. But I don't know if it's ok to recommend this change to all zsh users. I see the default `"terminal.integrated.profiles.osx.bash"` also has `"-l"` but this bad interaction doesn't happen. Is it a known issue, or it is WAI?
|
1.0
|
PATH mutation using EnvironmentVariableCollection prepend is overwritten in zsh - - VS Code Version: 1.80.1
- OS Version: Darwin x64 22.5.0
VS Code Go extension tries to change the integrated terminals' PATH environment variable using `EnvironmentVariableCollection.prepend` api.
I verified that the change contribution is known to vscode using "Terminal: Show Environment Contributions" command.
```
# Terminal Environment Changes
## Extension: golang.go
- `PATH=/Users/hakim/sdk/go1.20.3/bin:${env:PATH}`
## Extension: vscode.git
- `GIT_ASKPASS=/Applications/Visual Studio Code.app/Contents/Resources/app/extensions/git/dist/askpass.sh`
...
```
When I create a terminal tab with `bash`, I see the PATH change applied and go sdk path prepended as expected.
But, when I create a terminal with `zsh`, it looks like zsh prepends the login shell PATH again after applying our extension's PATH change.
Note all `"terminal.integrated.profiles.*"` settings were default.
A workaround I found is to remove `"-l"` arg from the default `"terminal.integrated.profiles.osx.zsh"` setting. But I don't know if it's ok to recommend this change to all zsh users. I see the default `"terminal.integrated.profiles.osx.bash"` also has `"-l"` but this bad interaction doesn't happen. Is it a known issue, or it is WAI?
|
process
|
path mutation using environmentvariablecollection prepend is overwritten in zsh vs code version os version darwin vs code go extension tries to change the integrated terminals path environment variable using environmentvariablecollection prepend api i verified that the change contribution is known to vscode using terminal show environment contributions command terminal environment changes extension golang go path users hakim sdk bin env path extension vscode git git askpass applications visual studio code app contents resources app extensions git dist askpass sh when i create a terminal tab with bash i see the path change applied and go sdk path prepended as expected but when i create a terminal with zsh it looks like zsh prepends the login shell path again after applying our extension s path change note all terminal integrated profiles settings were default a workaround i found is to remove l arg from the default terminal integrated profiles osx zsh setting but i don t know if it s ok to recommend this change to all zsh users i see the default terminal integrated profiles osx bash also has l but this bad interaction doesn t happen is it a known issue or it is wai
| 1
|
3,915
| 3,268,729,464
|
IssuesEvent
|
2015-10-23 13:12:05
|
mapbox/mapbox-gl-native
|
https://api.github.com/repos/mapbox/mapbox-gl-native
|
opened
|
Investigate better Android native debugging
|
Android build
|
In this Android BBQ video they provide some interesting tips on how to debug native code in an IDE via the emulator:
https://www.youtube.com/watch?v=IiAtqAWsPBY
/cc @kkaefer
|
1.0
|
Investigate better Android native debugging - In this Android BBQ video they provide some interesting tips on how to debug native code in an IDE via the emulator:
https://www.youtube.com/watch?v=IiAtqAWsPBY
/cc @kkaefer
|
non_process
|
investigate better android native debugging in this android bbq video they provide some interesting tips on how to debug native code in an ide via the emulator cc kkaefer
| 0
|
690,037
| 23,643,912,971
|
IssuesEvent
|
2022-08-25 19:54:08
|
IDAES/idaes-pse
|
https://api.github.com/repos/IDAES/idaes-pse
|
closed
|
Parameterize Henry's constant correlation
|
Priority:Normal property packages IDAES v2.0
|
Add parameters in Henry's constant N2O Analogy (Jiru et.al, 2012), in MEA_solvent.py.
The parameters are needed for the end-to-end UQ example, as they are estimated together with the concertation-based equilibrium constant parameters, using experimental VLE data from the literature.
|
1.0
|
Parameterize Henry's constant correlation - Add parameters in Henry's constant N2O Analogy (Jiru et.al, 2012), in MEA_solvent.py.
The parameters are needed for the end-to-end UQ example, as they are estimated together with the concertation-based equilibrium constant parameters, using experimental VLE data from the literature.
|
non_process
|
parameterize henry s constant correlation add parameters in henry s constant analogy jiru et al in mea solvent py the parameters are needed for the end to end uq example as they are estimated together with the concertation based equilibrium constant parameters using experimental vle data from the literature
| 0
|
2,717
| 5,581,212,215
|
IssuesEvent
|
2017-03-28 18:21:46
|
djspiewak/issue-testing
|
https://api.github.com/repos/djspiewak/issue-testing
|
opened
|
Find a workaround for the hamster
|
epic: Signup Process 2.0
|
We need the sacred hamster, but blood sacrifice is generally difficult for most prospective users. At least the vegans, anyway. Investigate solutions to completing signup in a more straightforward fashion.
|
1.0
|
Find a workaround for the hamster - We need the sacred hamster, but blood sacrifice is generally difficult for most prospective users. At least the vegans, anyway. Investigate solutions to completing signup in a more straightforward fashion.
|
process
|
find a workaround for the hamster we need the sacred hamster but blood sacrifice is generally difficult for most prospective users at least the vegans anyway investigate solutions to completing signup in a more straightforward fashion
| 1
|
57,162
| 3,081,245,130
|
IssuesEvent
|
2015-08-22 14:35:46
|
bitfighter/bitfighter
|
https://api.github.com/repos/bitfighter/bitfighter
|
closed
|
Bitfighter Crashes when hitting Esc in Host/Passwords
|
bug imported Priority-Medium
|
_From [corteocarl](https://code.google.com/u/corteocarl/) on February 05, 2014 18:55:11_
What steps will reproduce the problem? 1. Go to Bitfighter menu, then Host a Game
2. Go into the Passwords page
3. Hit escape What is the expected output? What do you see instead? It crashes :O What version of the product are you using? On what operating system? 019a 9366:2beb93cf9887 tip
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=386_
|
1.0
|
Bitfighter Crashes when hitting Esc in Host/Passwords - _From [corteocarl](https://code.google.com/u/corteocarl/) on February 05, 2014 18:55:11_
What steps will reproduce the problem? 1. Go to Bitfighter menu, then Host a Game
2. Go into the Passwords page
3. Hit escape What is the expected output? What do you see instead? It crashes :O What version of the product are you using? On what operating system? 019a 9366:2beb93cf9887 tip
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=386_
|
non_process
|
bitfighter crashes when hitting esc in host passwords from on february what steps will reproduce the problem go to bitfighter menu then host a game go into the passwords page hit escape what is the expected output what do you see instead it crashes o what version of the product are you using on what operating system tip original issue
| 0
|
12,755
| 15,113,430,882
|
IssuesEvent
|
2021-02-08 23:40:09
|
GoogleCloudPlatform/cloud-sql-jdbc-socket-factory
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-jdbc-socket-factory
|
closed
|
Create BOM for version management
|
type: process
|
Also investigate compile time warnings for incorrect versions of the connector-j.
|
1.0
|
Create BOM for version management - Also investigate compile time warnings for incorrect versions of the connector-j.
|
process
|
create bom for version management also investigate compile time warnings for incorrect versions of the connector j
| 1
|
5,523
| 8,381,047,613
|
IssuesEvent
|
2018-10-07 20:47:05
|
MichiganDataScienceTeam/googleanalytics
|
https://api.github.com/repos/MichiganDataScienceTeam/googleanalytics
|
opened
|
Preprocess: u'totals.bounces', u'totals.hits', u'totals.newVisits', u'totals.pageviews', u'totals.transactionRevenue', u'totals.visits',
|
easy preprocessing
|
Preprocess the following features:
u'totals.bounces',
u'totals.hits',
u'totals.newVisits',
u'totals.pageviews',
u'totals.transactionRevenue',
u'totals.visits',
1. Standardization: [http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling](http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling)
2. Impute missing values: [http://scikit-learn.org/stable/modules/impute.html](http://scikit-learn.org/stable/modules/impute.html)
3. Normalization: [http://scikit-learn.org/stable/modules/preprocessing.html#normalization](http://scikit-learn.org/stable/modules/preprocessing.html#normalization)
4. Encode categorical features (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features](http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features)
5. Discretization (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#discretization](http://scikit-learn.org/stable/modules/preprocessing.html#discretization)
[http://scikit-learn.org/stable/modules/preprocessing.html](http://scikit-learn.org/stable/modules/preprocessing.html)
|
1.0
|
Preprocess: u'totals.bounces', u'totals.hits', u'totals.newVisits', u'totals.pageviews', u'totals.transactionRevenue', u'totals.visits', - Preprocess the following features:
u'totals.bounces',
u'totals.hits',
u'totals.newVisits',
u'totals.pageviews',
u'totals.transactionRevenue',
u'totals.visits',
1. Standardization: [http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling](http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling)
2. Impute missing values: [http://scikit-learn.org/stable/modules/impute.html](http://scikit-learn.org/stable/modules/impute.html)
3. Normalization: [http://scikit-learn.org/stable/modules/preprocessing.html#normalization](http://scikit-learn.org/stable/modules/preprocessing.html#normalization)
4. Encode categorical features (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features](http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features)
5. Discretization (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#discretization](http://scikit-learn.org/stable/modules/preprocessing.html#discretization)
[http://scikit-learn.org/stable/modules/preprocessing.html](http://scikit-learn.org/stable/modules/preprocessing.html)
|
process
|
preprocess u totals bounces u totals hits u totals newvisits u totals pageviews u totals transactionrevenue u totals visits preprocess the following features u totals bounces u totals hits u totals newvisits u totals pageviews u totals transactionrevenue u totals visits standardization impute missing values normalization encode categorical features optional discretization optional
| 1
|
223,295
| 7,451,781,083
|
IssuesEvent
|
2018-03-29 05:16:01
|
pouladpld/Tank-Mania
|
https://api.github.com/repos/pouladpld/Tank-Mania
|
closed
|
Turn timeout
|
feature priority:high
|
### Feature
Each player will have X seconds to play. This timeout should be visible on the screen.
### Scene
All levels
### Description
- Define timeout period
- Implement timeout event on turn-changing mechanism
- A countdown-timer should be added to player screen
|
1.0
|
Turn timeout - ### Feature
Each player will have X seconds to play. This timeout should be visible on the screen.
### Scene
All levels
### Description
- Define timeout period
- Implement timeout event on turn-changing mechanism
- A countdown-timer should be added to player screen
|
non_process
|
turn timeout feature each player will have x seconds to play this timeout should be visible on the screen scene all levels description define timeout period implement timeout event on turn changing mechanism a countdown timer should be added to player screen
| 0
|
9,456
| 12,438,434,539
|
IssuesEvent
|
2020-05-26 08:25:13
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
reopened
|
Document Prisma Client as a shared private dependency
|
kind/docs process/candidate
|
## Problem
It can be useful to share generated database clients across projects, maybe you're got many applications talking to the same database.
## Suggested Solution
Document this workflow: https://github.com/prisma/prisma/issues/1787#issuecomment-633881873
Based on: https://github.com/prisma/prisma/issues/458 https://github.com/prisma/prisma/issues/1787
## Alternative Solution
Test and Document:
- [git submodules](https://git-scm.com/book/en/v2/Git-Tools-Submodules)
- [git subtree](https://git.kernel.org/pub/scm/git/git.git/plain/contrib/subtree/git-subtree.txt)
- [npm private repo](https://docs.npmjs.com/creating-and-publishing-private-packages)
|
1.0
|
Document Prisma Client as a shared private dependency - ## Problem
It can be useful to share generated database clients across projects, maybe you're got many applications talking to the same database.
## Suggested Solution
Document this workflow: https://github.com/prisma/prisma/issues/1787#issuecomment-633881873
Based on: https://github.com/prisma/prisma/issues/458 https://github.com/prisma/prisma/issues/1787
## Alternative Solution
Test and Document:
- [git submodules](https://git-scm.com/book/en/v2/Git-Tools-Submodules)
- [git subtree](https://git.kernel.org/pub/scm/git/git.git/plain/contrib/subtree/git-subtree.txt)
- [npm private repo](https://docs.npmjs.com/creating-and-publishing-private-packages)
|
process
|
document prisma client as a shared private dependency problem it can be useful to share generated database clients across projects maybe you re got many applications talking to the same database suggested solution document this workflow based on alternative solution test and document
| 1
|
17,866
| 23,812,912,445
|
IssuesEvent
|
2022-09-05 01:12:38
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Mosquito from Popcorn (1991)
|
suggested title in process
|
Title: Mosquito
Type (film/tv show): Film
Film or show in which it appears: Popcorn (1991)
Is the parent film/show streaming anywhere? Yes (Shudder and AMC+); https://www.justwatch.com/us/movie/popcorn-1991
About when in the parent film/show does it appear? 0:26:35
Actual footage of the film/show can be seen (yes/no)? yes
Starring
Randall Todd
Alison Holt
Introducing
Cubby Scott ad "Corky"
Quotes:
Send the Air Force, Mr. President. Don't forget the Army, Navy, and the Coast Guard.
We need every plane, tank, and battleship you've got, Mr. President.
This is one hell of a bug.
- Nothing fazes it. We'll have to drop the A-bomb.
- No, Jim, that's what started all this trouble in the first place.
Screenshots:













|
1.0
|
Add Mosquito from Popcorn (1991) - Title: Mosquito
Type (film/tv show): Film
Film or show in which it appears: Popcorn (1991)
Is the parent film/show streaming anywhere? Yes (Shudder and AMC+); https://www.justwatch.com/us/movie/popcorn-1991
About when in the parent film/show does it appear? 0:26:35
Actual footage of the film/show can be seen (yes/no)? yes
Starring
Randall Todd
Alison Holt
Introducing
Cubby Scott ad "Corky"
Quotes:
Send the Air Force, Mr. President. Don't forget the Army, Navy, and the Coast Guard.
We need every plane, tank, and battleship you've got, Mr. President.
This is one hell of a bug.
- Nothing fazes it. We'll have to drop the A-bomb.
- No, Jim, that's what started all this trouble in the first place.
Screenshots:













|
process
|
add mosquito from popcorn title mosquito type film tv show film film or show in which it appears popcorn is the parent film show streaming anywhere yes shudder and amc about when in the parent film show does it appear actual footage of the film show can be seen yes no yes starring randall todd alison holt introducing cubby scott ad corky quotes send the air force mr president don t forget the army navy and the coast guard we need every plane tank and battleship you ve got mr president this is one hell of a bug nothing fazes it we ll have to drop the a bomb no jim that s what started all this trouble in the first place screenshots
| 1
|
374,109
| 11,072,114,922
|
IssuesEvent
|
2019-12-12 09:40:08
|
AbsaOSS/hyperdrive-trigger
|
https://api.github.com/repos/AbsaOSS/hyperdrive-trigger
|
opened
|
DB connection pool - multiple pools are created
|
backend bug priority: low
|
DB connection pool - multiple pools are created.
For each repository implementation, a new thread pool is created.
All connections to DB should be managed by one thread pool.
|
1.0
|
DB connection pool - multiple pools are created - DB connection pool - multiple pools are created.
For each repository implementation, a new thread pool is created.
All connections to DB should be managed by one thread pool.
|
non_process
|
db connection pool multiple pools are created db connection pool multiple pools are created for each repository implementation a new thread pool is created all connections to db should be managed by one thread pool
| 0
|
10,679
| 13,462,851,065
|
IssuesEvent
|
2020-09-09 16:40:07
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Merge `PrismaQueryEngineError` into `PrismaClientUnknownRequestError`
|
kind/improvement process/candidate team/typescript
|
The `PrismaQueryEngineError` could just be a prisma client request error.
They're just used here: https://github.com/prisma/prisma/blob/master/src/packages/engine-core/src/client.ts#L57
|
1.0
|
Merge `PrismaQueryEngineError` into `PrismaClientUnknownRequestError` - The `PrismaQueryEngineError` could just be a prisma client request error.
They're just used here: https://github.com/prisma/prisma/blob/master/src/packages/engine-core/src/client.ts#L57
|
process
|
merge prismaqueryengineerror into prismaclientunknownrequesterror the prismaqueryengineerror could just be a prisma client request error they re just used here
| 1
|
20,643
| 27,321,239,777
|
IssuesEvent
|
2023-02-24 20:03:19
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Pivot Table query optimization
|
Type:Bug Priority:P2 .Performance Querying/Processor Querying/Nested Queries Difficulty:Hard .Backend Visualization/Tables
|
@luizarakaki comment:
We should improve the Pivot Table query in two ways:
- [ ] Move filtering clauses to the inner query (so it actually use indexes).
- [ ] Select only relevant fields in the inner query. If the table has more fields than used for breakouts, we should get only the necessary to build the pivot.
---
_Original issue_
Hello metabase team,
I'm playing with pivot table and found them extremely slow.
After some digging, I'm not alone :
- https://discourse.metabase.com/t/pivot-tables-extremely-slow/13694
- https://discourse.metabase.com/t/how-to-improve-performance-in-data-present-in-pivod-table/16350/8
- https://discourse.metabase.com/t/pivot-table-is-just-running/16602
I did some research on how table are pivoted looking at queries generated on my database and I probably found an optimization with filters.
**Example source query :**
```
SELECT `REQUEST`.`REQ_ID` AS `REQ_ID`,
`REQUEST`.`REQ_DATETIME` AS `REQ_DATETIME`,
`Agence`.`AG_NOM` AS `Agence__AG_NOM`,
`Marque`.`LIBELLE` AS `Marque__LIBELLE`,
`Status`.`STA_NAME` AS `Status__STA_NAME`
FROM `REQUEST`
LEFT JOIN `AGENCE` `Agence` ON `REQUEST`.`AG_ID` = `Agence`.`AG_ID`
LEFT JOIN `REF_SOCIETE` `Marque` ON `Agence`.`SOC_ID` = `Marque`.`SOC_ID`
LEFT JOIN `STATUS` `Status` ON `REQUEST`.`REQ_LAST_STATUS` = `Status`.`STA_ID`
LEFT JOIN `AGENCE` `AGENCE__via__AG_ID` ON `REQUEST`.`AG_ID` = `AGENCE__via__AG_ID`.`AG_ID`
WHERE (`REQUEST`.`REQ_DATETIME` >= date(now(6))
AND `REQUEST`.`REQ_DATETIME` < date(date_add(now(6), INTERVAL 1 DAY)))
```
Result : 145 row(s) fetched - 39ms (4ms fetch), on avr. 15, 09:39:30
**Pivoted one :**
```
SELECT `source`.`Marque__LIBELLE` AS `Marque__LIBELLE`,
`source`.`Agence__AG_NOM` AS `Agence__AG_NOM`,
`source`.`Status__STA_NAME` AS `Status__STA_NAME`,
`source`.`pivot-grouping` AS `pivot-grouping`,
count(DISTINCT `source`.`REQ_ID`) AS `count`
FROM
(SELECT `REQUEST`.`REQ_ID` AS `REQ_ID`,
`REQUEST`.`REQ_DATETIME` AS `REQ_DATETIME`,
abs(0) AS `pivot-grouping`,
`Agence`.`AG_NOM` AS `Agence__AG_NOM`,
`Marque`.`LIBELLE` AS `Marque__LIBELLE`,
`Status`.`STA_NAME` AS `Status__STA_NAME`
FROM `REQUEST`
LEFT JOIN `AGENCE` `Agence` ON `REQUEST`.`AG_ID` = `Agence`.`AG_ID`
LEFT JOIN `REF_SOCIETE` `Marque` ON `Agence`.`SOC_ID` = `Marque`.`SOC_ID`
LEFT JOIN `STATUS` `Status` ON `REQUEST`.`REQ_LAST_STATUS` = `Status`.`STA_ID`
LEFT JOIN `AGENCE` `AGENCE__via__AG_ID` ON `REQUEST`.`AG_ID` = `AGENCE__via__AG_ID`.`AG_ID`) `source`
WHERE (`source`.`REQ_DATETIME` >= date(now(6))
AND `source`.`REQ_DATETIME` < date(date_add(now(6), INTERVAL 1 DAY)))
GROUP BY `source`.`Marque__LIBELLE`,
`source`.`Agence__AG_NOM`,
`source`.`Status__STA_NAME`,
`source`.`pivot-grouping`
ORDER BY `source`.`Marque__LIBELLE` ASC,
`source`.`Agence__AG_NOM` ASC,
`source`.`Status__STA_NAME` ASC,
`source`.`pivot-grouping` ASC
```
Result : 16 row(s) fetched - 10.524s (1ms fetch), on avr. 15, 09:45:33
Problem : the "source" sub query is not filtered as asked in initial query, only the pivot is filtered.
**Optimized pivot** (where clause on source, not on pivot) :
```
SELECT `source`.`Marque__LIBELLE` AS `Marque__LIBELLE`,
`source`.`Agence__AG_NOM` AS `Agence__AG_NOM`,
`source`.`Status__STA_NAME` AS `Status__STA_NAME`,
`source`.`pivot-grouping` AS `pivot-grouping`,
count(DISTINCT `source`.`REQ_ID`) AS `count`
FROM
(SELECT `REQUEST`.`REQ_ID` AS `REQ_ID`,
`REQUEST`.`REQ_DATETIME` AS `REQ_DATETIME`,
abs(0) AS `pivot-grouping`,
`Agence`.`AG_NOM` AS `Agence__AG_NOM`,
`Marque`.`LIBELLE` AS `Marque__LIBELLE`,
`Status`.`STA_NAME` AS `Status__STA_NAME`
FROM `REQUEST`
LEFT JOIN `AGENCE` `Agence` ON `REQUEST`.`AG_ID` = `Agence`.`AG_ID`
LEFT JOIN `REF_SOCIETE` `Marque` ON `Agence`.`SOC_ID` = `Marque`.`SOC_ID`
LEFT JOIN `STATUS` `Status` ON `REQUEST`.`REQ_LAST_STATUS` = `Status`.`STA_ID`
LEFT JOIN `AGENCE` `AGENCE__via__AG_ID` ON `REQUEST`.`AG_ID` = `AGENCE__via__AG_ID`.`AG_ID`
WHERE (`REQUEST`.`REQ_DATETIME` >= date(now(6))
AND `REQUEST`.`REQ_DATETIME` < date(date_add(now(6), INTERVAL 1 DAY))) ) `source`
GROUP BY `source`.`Marque__LIBELLE`,
`source`.`Agence__AG_NOM`,
`source`.`Status__STA_NAME`,
`source`.`pivot-grouping`
ORDER BY `source`.`Marque__LIBELLE` ASC,
`source`.`Agence__AG_NOM` ASC,
`source`.`Status__STA_NAME` ASC,
`source`.`pivot-grouping` ASC
```
16 row(s) fetched - 20ms, on avr. 15, 09:47:26
From 10.524s to 20ms, it's 99.8% better.
Do you think this can be implemented ?
:arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment
|
1.0
|
Pivot Table query optimization - @luizarakaki comment:
We should improve the Pivot Table query in two ways:
- [ ] Move filtering clauses to the inner query (so it actually use indexes).
- [ ] Select only relevant fields in the inner query. If the table has more fields than used for breakouts, we should get only the necessary to build the pivot.
---
_Original issue_
Hello metabase team,
I'm playing with pivot table and found them extremely slow.
After some digging, I'm not alone :
- https://discourse.metabase.com/t/pivot-tables-extremely-slow/13694
- https://discourse.metabase.com/t/how-to-improve-performance-in-data-present-in-pivod-table/16350/8
- https://discourse.metabase.com/t/pivot-table-is-just-running/16602
I did some research on how table are pivoted looking at queries generated on my database and I probably found an optimization with filters.
**Example source query :**
```
SELECT `REQUEST`.`REQ_ID` AS `REQ_ID`,
`REQUEST`.`REQ_DATETIME` AS `REQ_DATETIME`,
`Agence`.`AG_NOM` AS `Agence__AG_NOM`,
`Marque`.`LIBELLE` AS `Marque__LIBELLE`,
`Status`.`STA_NAME` AS `Status__STA_NAME`
FROM `REQUEST`
LEFT JOIN `AGENCE` `Agence` ON `REQUEST`.`AG_ID` = `Agence`.`AG_ID`
LEFT JOIN `REF_SOCIETE` `Marque` ON `Agence`.`SOC_ID` = `Marque`.`SOC_ID`
LEFT JOIN `STATUS` `Status` ON `REQUEST`.`REQ_LAST_STATUS` = `Status`.`STA_ID`
LEFT JOIN `AGENCE` `AGENCE__via__AG_ID` ON `REQUEST`.`AG_ID` = `AGENCE__via__AG_ID`.`AG_ID`
WHERE (`REQUEST`.`REQ_DATETIME` >= date(now(6))
AND `REQUEST`.`REQ_DATETIME` < date(date_add(now(6), INTERVAL 1 DAY)))
```
Result : 145 row(s) fetched - 39ms (4ms fetch), on avr. 15, 09:39:30
**Pivoted one :**
```
SELECT `source`.`Marque__LIBELLE` AS `Marque__LIBELLE`,
`source`.`Agence__AG_NOM` AS `Agence__AG_NOM`,
`source`.`Status__STA_NAME` AS `Status__STA_NAME`,
`source`.`pivot-grouping` AS `pivot-grouping`,
count(DISTINCT `source`.`REQ_ID`) AS `count`
FROM
(SELECT `REQUEST`.`REQ_ID` AS `REQ_ID`,
`REQUEST`.`REQ_DATETIME` AS `REQ_DATETIME`,
abs(0) AS `pivot-grouping`,
`Agence`.`AG_NOM` AS `Agence__AG_NOM`,
`Marque`.`LIBELLE` AS `Marque__LIBELLE`,
`Status`.`STA_NAME` AS `Status__STA_NAME`
FROM `REQUEST`
LEFT JOIN `AGENCE` `Agence` ON `REQUEST`.`AG_ID` = `Agence`.`AG_ID`
LEFT JOIN `REF_SOCIETE` `Marque` ON `Agence`.`SOC_ID` = `Marque`.`SOC_ID`
LEFT JOIN `STATUS` `Status` ON `REQUEST`.`REQ_LAST_STATUS` = `Status`.`STA_ID`
LEFT JOIN `AGENCE` `AGENCE__via__AG_ID` ON `REQUEST`.`AG_ID` = `AGENCE__via__AG_ID`.`AG_ID`) `source`
WHERE (`source`.`REQ_DATETIME` >= date(now(6))
AND `source`.`REQ_DATETIME` < date(date_add(now(6), INTERVAL 1 DAY)))
GROUP BY `source`.`Marque__LIBELLE`,
`source`.`Agence__AG_NOM`,
`source`.`Status__STA_NAME`,
`source`.`pivot-grouping`
ORDER BY `source`.`Marque__LIBELLE` ASC,
`source`.`Agence__AG_NOM` ASC,
`source`.`Status__STA_NAME` ASC,
`source`.`pivot-grouping` ASC
```
Result : 16 row(s) fetched - 10.524s (1ms fetch), on avr. 15, 09:45:33
Problem : the "source" sub query is not filtered as asked in initial query, only the pivot is filtered.
**Optimized pivot** (where clause on source, not on pivot) :
```
SELECT `source`.`Marque__LIBELLE` AS `Marque__LIBELLE`,
`source`.`Agence__AG_NOM` AS `Agence__AG_NOM`,
`source`.`Status__STA_NAME` AS `Status__STA_NAME`,
`source`.`pivot-grouping` AS `pivot-grouping`,
count(DISTINCT `source`.`REQ_ID`) AS `count`
FROM
(SELECT `REQUEST`.`REQ_ID` AS `REQ_ID`,
`REQUEST`.`REQ_DATETIME` AS `REQ_DATETIME`,
abs(0) AS `pivot-grouping`,
`Agence`.`AG_NOM` AS `Agence__AG_NOM`,
`Marque`.`LIBELLE` AS `Marque__LIBELLE`,
`Status`.`STA_NAME` AS `Status__STA_NAME`
FROM `REQUEST`
LEFT JOIN `AGENCE` `Agence` ON `REQUEST`.`AG_ID` = `Agence`.`AG_ID`
LEFT JOIN `REF_SOCIETE` `Marque` ON `Agence`.`SOC_ID` = `Marque`.`SOC_ID`
LEFT JOIN `STATUS` `Status` ON `REQUEST`.`REQ_LAST_STATUS` = `Status`.`STA_ID`
LEFT JOIN `AGENCE` `AGENCE__via__AG_ID` ON `REQUEST`.`AG_ID` = `AGENCE__via__AG_ID`.`AG_ID`
WHERE (`REQUEST`.`REQ_DATETIME` >= date(now(6))
AND `REQUEST`.`REQ_DATETIME` < date(date_add(now(6), INTERVAL 1 DAY))) ) `source`
GROUP BY `source`.`Marque__LIBELLE`,
`source`.`Agence__AG_NOM`,
`source`.`Status__STA_NAME`,
`source`.`pivot-grouping`
ORDER BY `source`.`Marque__LIBELLE` ASC,
`source`.`Agence__AG_NOM` ASC,
`source`.`Status__STA_NAME` ASC,
`source`.`pivot-grouping` ASC
```
16 row(s) fetched - 20ms, on avr. 15, 09:47:26
From 10.524s to 20ms, it's 99.8% better.
Do you think this can be implemented ?
:arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment
|
process
|
pivot table query optimization luizarakaki comment we should improve the pivot table query in two ways move filtering clauses to the inner query so it actually use indexes select only relevant fields in the inner query if the table has more fields than used for breakouts we should get only the necessary to build the pivot original issue hello metabase team i m playing with pivot table and found them extremely slow after some digging i m not alone i did some research on how table are pivoted looking at queries generated on my database and i probably found an optimization with filters example source query select request req id as req id request req datetime as req datetime agence ag nom as agence ag nom marque libelle as marque libelle status sta name as status sta name from request left join agence agence on request ag id agence ag id left join ref societe marque on agence soc id marque soc id left join status status on request req last status status sta id left join agence agence via ag id on request ag id agence via ag id ag id where request req datetime date now and request req datetime date date add now interval day result row s fetched fetch on avr pivoted one select source marque libelle as marque libelle source agence ag nom as agence ag nom source status sta name as status sta name source pivot grouping as pivot grouping count distinct source req id as count from select request req id as req id request req datetime as req datetime abs as pivot grouping agence ag nom as agence ag nom marque libelle as marque libelle status sta name as status sta name from request left join agence agence on request ag id agence ag id left join ref societe marque on agence soc id marque soc id left join status status on request req last status status sta id left join agence agence via ag id on request ag id agence via ag id ag id source where source req datetime date now and source req datetime date date add now interval day group by source marque libelle source agence ag nom source status sta name source pivot grouping order by source marque libelle asc source agence ag nom asc source status sta name asc source pivot grouping asc result row s fetched fetch on avr problem the source sub query is not filtered as asked in initial query only the pivot is filtered optimized pivot where clause on source not on pivot select source marque libelle as marque libelle source agence ag nom as agence ag nom source status sta name as status sta name source pivot grouping as pivot grouping count distinct source req id as count from select request req id as req id request req datetime as req datetime abs as pivot grouping agence ag nom as agence ag nom marque libelle as marque libelle status sta name as status sta name from request left join agence agence on request ag id agence ag id left join ref societe marque on agence soc id marque soc id left join status status on request req last status status sta id left join agence agence via ag id on request ag id agence via ag id ag id where request req datetime date now and request req datetime date date add now interval day source group by source marque libelle source agence ag nom source status sta name source pivot grouping order by source marque libelle asc source agence ag nom asc source status sta name asc source pivot grouping asc row s fetched on avr from to it s better do you think this can be implemented arrow down please click the reaction instead of leaving a or update comment
| 1
|
180,462
| 30,507,789,067
|
IssuesEvent
|
2023-07-18 18:15:48
|
carbon-design-system/ibm-products
|
https://api.github.com/repos/carbon-design-system/ibm-products
|
closed
|
Datagrid design review: Empty states
|
status: ready for design review component: Datagrid v2 v1
|
## Design review
- Component epic #1607 #2499
- Links: [PAL guidance](https://pages.github.ibm.com/cdai-design/pal/components/data-table/overview) / [Storybook v2](https://ibm-products.carbondesignsystem.com/?path=/story/ibm-products-components-datagrid-datagrid-canary--empty-state) / [Storybook v1](https://v1-ibm-products.carbondesignsystem.com/?path=/story/ibm-products-components-datagrid-datagrid-canary--empty-state)
- [Empty state](https://pages.github.ibm.com/cdai-design/pal/patterns/empty-state/usage)
---
### Standards
- [ ] All pattern updates/changes/iterations have been discussed with the
component developer
- [ ] Patterns have been approved by either DSAG or another approving entity
### Pattern and behavior
- [ ] The component behaves according to the guidelines set by the pattern
maintainer
- [ ] The componentβs UI matches the pattern specifications set by the pattern
maintainer
- [ ] The componentβs motion matches the specifications set by the pattern
maintainer(s)
- [ ] The UI produced is responsive, cross-browser, and responds to the
currently set Carbon theme.
- [ ] Colors, themes, sizes and additional props are true and correct, aligning
with Carbon set tokens (unless otherwise specified by the pattern)
- [ ] Paddings/margins/spacings are true and correct, in both desktop and mobile
views
### Storybook
- [ ] A functioning component renders in Storybook
- [ ] The Storybook displays multiple scenarios that show how the component can
be used
- [ ] Changing props in the Storybook updates the component
|
1.0
|
Datagrid design review: Empty states - ## Design review
- Component epic #1607 #2499
- Links: [PAL guidance](https://pages.github.ibm.com/cdai-design/pal/components/data-table/overview) / [Storybook v2](https://ibm-products.carbondesignsystem.com/?path=/story/ibm-products-components-datagrid-datagrid-canary--empty-state) / [Storybook v1](https://v1-ibm-products.carbondesignsystem.com/?path=/story/ibm-products-components-datagrid-datagrid-canary--empty-state)
- [Empty state](https://pages.github.ibm.com/cdai-design/pal/patterns/empty-state/usage)
---
### Standards
- [ ] All pattern updates/changes/iterations have been discussed with the
component developer
- [ ] Patterns have been approved by either DSAG or another approving entity
### Pattern and behavior
- [ ] The component behaves according to the guidelines set by the pattern
maintainer
- [ ] The componentβs UI matches the pattern specifications set by the pattern
maintainer
- [ ] The componentβs motion matches the specifications set by the pattern
maintainer(s)
- [ ] The UI produced is responsive, cross-browser, and responds to the
currently set Carbon theme.
- [ ] Colors, themes, sizes and additional props are true and correct, aligning
with Carbon set tokens (unless otherwise specified by the pattern)
- [ ] Paddings/margins/spacings are true and correct, in both desktop and mobile
views
### Storybook
- [ ] A functioning component renders in Storybook
- [ ] The Storybook displays multiple scenarios that show how the component can
be used
- [ ] Changing props in the Storybook updates the component
|
non_process
|
datagrid design review empty states design review component epic links standards all pattern updates changes iterations have been discussed with the component developer patterns have been approved by either dsag or another approving entity pattern and behavior the component behaves according to the guidelines set by the pattern maintainer the componentβs ui matches the pattern specifications set by the pattern maintainer the componentβs motion matches the specifications set by the pattern maintainer s the ui produced is responsive cross browser and responds to the currently set carbon theme colors themes sizes and additional props are true and correct aligning with carbon set tokens unless otherwise specified by the pattern paddings margins spacings are true and correct in both desktop and mobile views storybook a functioning component renders in storybook the storybook displays multiple scenarios that show how the component can be used changing props in the storybook updates the component
| 0
|
655,514
| 21,693,470,448
|
IssuesEvent
|
2022-05-09 17:35:38
|
googleapis/java-vision
|
https://api.github.com/repos/googleapis/java-vision
|
closed
|
vision.it.ITSystemTest: detectTextTest failed
|
:rotating_light: priority: p1 type: bug api: vision flakybot: issue flakybot: flaky
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: ac6eb364f91b9ece1e975594f5288409de432f63
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1509fa3c-3733-4fa3-bfa2-e7ed36cb59db), [Sponge](http://sponge2/1509fa3c-3733-4fa3-bfa2-e7ed36cb59db)
status: failed
<details><summary>Test output</summary><br><pre>expected to contain:
37%
but was:
[System Software Update
Preparing to install...
After preparation is complete, the PS4 will automatically restart and the update file will be
installed.
37%
Back
gus class, System, Software, Update, Preparing, to, install, ..., After, preparation, is, complete, ,, the, PS4, will, automatically, restart, and, the, update, file, will, be, installed, ., 37, %, Back, gus, class]
at com.google.cloud.vision.it.ITSystemTest.detectTextTest(ITSystemTest.java:396)
</pre></details>
|
1.0
|
vision.it.ITSystemTest: detectTextTest failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: ac6eb364f91b9ece1e975594f5288409de432f63
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1509fa3c-3733-4fa3-bfa2-e7ed36cb59db), [Sponge](http://sponge2/1509fa3c-3733-4fa3-bfa2-e7ed36cb59db)
status: failed
<details><summary>Test output</summary><br><pre>expected to contain:
37%
but was:
[System Software Update
Preparing to install...
After preparation is complete, the PS4 will automatically restart and the update file will be
installed.
37%
Back
gus class, System, Software, Update, Preparing, to, install, ..., After, preparation, is, complete, ,, the, PS4, will, automatically, restart, and, the, update, file, will, be, installed, ., 37, %, Back, gus, class]
at com.google.cloud.vision.it.ITSystemTest.detectTextTest(ITSystemTest.java:396)
</pre></details>
|
non_process
|
vision it itsystemtest detecttexttest failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output expected to contain but was system software update preparing to install after preparation is complete the will automatically restart and the update file will be installed back gus class system software update preparing to install after preparation is complete the will automatically restart and the update file will be installed back gus class at com google cloud vision it itsystemtest detecttexttest itsystemtest java
| 0
|
21,592
| 10,666,988,587
|
IssuesEvent
|
2019-10-19 08:45:59
|
tomdgl397/goof
|
https://api.github.com/repos/tomdgl397/goof
|
opened
|
CVE-2019-17426 (Medium) detected in mongoose-5.4.10.tgz
|
security vulnerability
|
## CVE-2019-17426 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongoose-5.4.10.tgz</b></p></summary>
<p>Mongoose MongoDB ODM</p>
<p>Library home page: <a href="https://registry.npmjs.org/mongoose/-/mongoose-5.4.10.tgz">https://registry.npmjs.org/mongoose/-/mongoose-5.4.10.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/goof/package.json</p>
<p>Path to vulnerable library: /goof/node_modules/mongoose/package.json</p>
<p>
Dependency Hierarchy:
- :x: **mongoose-5.4.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tomdgl397/goof/commit/6240bab6415a7ccc6742d55c66236b479ec50ae6">6240bab6415a7ccc6742d55c66236b479ec50ae6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Automattic Mongoose through 5.7.4 allows attackers to bypass access control (in some applications) because any query object with a _bsontype attribute is ignored. For example, adding "_bsontype":"a" can sometimes interfere with a query filter. NOTE: this CVE is about Mongoose's failure to work around this _bsontype special case that exists in older versions of the bson parser (aka the mongodb/js-bson project).
<p>Publish Date: 2019-10-10
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17426>CVE-2019-17426</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.4</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-17426 (Medium) detected in mongoose-5.4.10.tgz - ## CVE-2019-17426 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongoose-5.4.10.tgz</b></p></summary>
<p>Mongoose MongoDB ODM</p>
<p>Library home page: <a href="https://registry.npmjs.org/mongoose/-/mongoose-5.4.10.tgz">https://registry.npmjs.org/mongoose/-/mongoose-5.4.10.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/goof/package.json</p>
<p>Path to vulnerable library: /goof/node_modules/mongoose/package.json</p>
<p>
Dependency Hierarchy:
- :x: **mongoose-5.4.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tomdgl397/goof/commit/6240bab6415a7ccc6742d55c66236b479ec50ae6">6240bab6415a7ccc6742d55c66236b479ec50ae6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Automattic Mongoose through 5.7.4 allows attackers to bypass access control (in some applications) because any query object with a _bsontype attribute is ignored. For example, adding "_bsontype":"a" can sometimes interfere with a query filter. NOTE: this CVE is about Mongoose's failure to work around this _bsontype special case that exists in older versions of the bson parser (aka the mongodb/js-bson project).
<p>Publish Date: 2019-10-10
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17426>CVE-2019-17426</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.4</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in mongoose tgz cve medium severity vulnerability vulnerable library mongoose tgz mongoose mongodb odm library home page a href path to dependency file tmp ws scm goof package json path to vulnerable library goof node modules mongoose package json dependency hierarchy x mongoose tgz vulnerable library found in head commit a href vulnerability details automattic mongoose through allows attackers to bypass access control in some applications because any query object with a bsontype attribute is ignored for example adding bsontype a can sometimes interfere with a query filter note this cve is about mongoose s failure to work around this bsontype special case that exists in older versions of the bson parser aka the mongodb js bson project publish date url a href cvss score details base score metrics not available step up your open source security game with whitesource
| 0
|
172,008
| 21,031,000,065
|
IssuesEvent
|
2022-03-31 01:00:13
|
EliyaC/SecurityShepherd
|
https://api.github.com/repos/EliyaC/SecurityShepherd
|
opened
|
CVE-2022-22950 (Medium) detected in spring-expression-5.1.1.RELEASE.jar
|
security vulnerability
|
## CVE-2022-22950 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.1.1.RELEASE.jar</b></p></summary>
<p>Spring Expression Language (SpEL)</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-expression/5.1.1.RELEASE/spring-expression-5.1.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-context-5.1.1.RELEASE.jar (Root Library)
- :x: **spring-expression-5.1.1.RELEASE.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition
<p>Publish Date: 2022-01-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22950>CVE-2022-22950</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22950">https://tanzu.vmware.com/security/cve-2022-22950</a></p>
<p>Release Date: 2022-01-11</p>
<p>Fix Resolution: org.springframework:spring-expression:5.3.17</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-expression","packageVersion":"5.1.1.RELEASE","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework:spring-context:5.1.1.RELEASE;org.springframework:spring-expression:5.1.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-expression:5.3.17","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2022-22950","vulnerabilityDetails":"In Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22950","cvss3Severity":"medium","cvss3Score":"5.4","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2022-22950 (Medium) detected in spring-expression-5.1.1.RELEASE.jar - ## CVE-2022-22950 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.1.1.RELEASE.jar</b></p></summary>
<p>Spring Expression Language (SpEL)</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-expression/5.1.1.RELEASE/spring-expression-5.1.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-context-5.1.1.RELEASE.jar (Root Library)
- :x: **spring-expression-5.1.1.RELEASE.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition
<p>Publish Date: 2022-01-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22950>CVE-2022-22950</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22950">https://tanzu.vmware.com/security/cve-2022-22950</a></p>
<p>Release Date: 2022-01-11</p>
<p>Fix Resolution: org.springframework:spring-expression:5.3.17</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-expression","packageVersion":"5.1.1.RELEASE","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework:spring-context:5.1.1.RELEASE;org.springframework:spring-expression:5.1.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-expression:5.3.17","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2022-22950","vulnerabilityDetails":"In Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22950","cvss3Severity":"medium","cvss3Score":"5.4","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in spring expression release jar cve medium severity vulnerability vulnerable library spring expression release jar spring expression language spel library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org springframework spring expression release spring expression release jar dependency hierarchy spring context release jar root library x spring expression release jar vulnerable library found in base branch dev vulnerability details in spring framework versions and older unsupported versions it is possible for a user to provide a specially crafted spel expression that may cause a denial of service condition publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring expression isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework spring context release org springframework spring expression release isminimumfixversionavailable true minimumfixversion org springframework spring expression isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in spring framework versions and older unsupported versions it is possible for a user to provide a specially crafted spel expression that may cause a denial of service condition vulnerabilityurl
| 0
|
22,024
| 30,536,010,523
|
IssuesEvent
|
2023-07-19 17:24:18
|
USGS-WiM/StreamStats
|
https://api.github.com/repos/USGS-WiM/StreamStats
|
closed
|
Add upload file name to Batch Status and Manage Queue tabs
|
Batch Processor
|
In the Batch Status and Manage Queue tabs, add a column to display the name of the file that the user uploaded for their batch.
|
1.0
|
Add upload file name to Batch Status and Manage Queue tabs - In the Batch Status and Manage Queue tabs, add a column to display the name of the file that the user uploaded for their batch.
|
process
|
add upload file name to batch status and manage queue tabs in the batch status and manage queue tabs add a column to display the name of the file that the user uploaded for their batch
| 1
|
294,172
| 9,014,048,312
|
IssuesEvent
|
2019-02-05 21:14:26
|
AugurProject/augur
|
https://api.github.com/repos/AugurProject/augur
|
closed
|
log removals for market state changes
|
Chore Priority: Low
|
marketState: awaiting_net_window
if it is rolled back
should be
marketState: awiating_designated_report, removed: true (more like, this is the new state, due to a removal)
...
@epheph is this task still needed or desired?
-bthaile
...
Technically, yes. It's a very subtle issue that could happen under rare circumstances: If a new fee window happens on a block that gets re-org'd. we will move the things too soon to their new state, and the block that is re-org'd to might not have a timestamp that moves the fee window forward until the NEXT block. The impact is minor and unlikely, but still is a flaw
-epheph
...
Look into if this is still true.
-epheph
|
1.0
|
log removals for market state changes - marketState: awaiting_net_window
if it is rolled back
should be
marketState: awiating_designated_report, removed: true (more like, this is the new state, due to a removal)
...
@epheph is this task still needed or desired?
-bthaile
...
Technically, yes. It's a very subtle issue that could happen under rare circumstances: If a new fee window happens on a block that gets re-org'd. we will move the things too soon to their new state, and the block that is re-org'd to might not have a timestamp that moves the fee window forward until the NEXT block. The impact is minor and unlikely, but still is a flaw
-epheph
...
Look into if this is still true.
-epheph
|
non_process
|
log removals for market state changes marketstate awaiting net window if it is rolled back should be marketstate awiating designated report removed true more like this is the new state due to a removal epheph is this task still needed or desired bthaile technically yes it s a very subtle issue that could happen under rare circumstances if a new fee window happens on a block that gets re org d we will move the things too soon to their new state and the block that is re org d to might not have a timestamp that moves the fee window forward until the next block the impact is minor and unlikely but still is a flaw epheph look into if this is still true epheph
| 0
|
22,054
| 30,572,559,066
|
IssuesEvent
|
2023-07-21 00:27:30
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
roblox-pyc 1.19.74 has 2 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.19.74",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.19.74/src/robloxpy.py:134",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.19.74/src/robloxpy.py:141",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpbednpyhh/roblox-pyc"
}
}```
|
1.0
|
roblox-pyc 1.19.74 has 2 GuardDog issues - https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.19.74",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.19.74/src/robloxpy.py:134",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.19.74/src/robloxpy.py:141",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpbednpyhh/roblox-pyc"
}
}```
|
process
|
roblox pyc has guarddog issues dependency roblox pyc version result issues errors results silent process execution location roblox pyc src robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc src robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpbednpyhh roblox pyc
| 1
|
13,470
| 15,955,600,312
|
IssuesEvent
|
2021-04-15 14:46:21
|
sigstore/rekor
|
https://api.github.com/repos/sigstore/rekor
|
closed
|
pin releases in k8s config
|
release-process
|
pre-release step: review all k8s configs and ensure we use tags/digests where appropriate.
|
1.0
|
pin releases in k8s config - pre-release step: review all k8s configs and ensure we use tags/digests where appropriate.
|
process
|
pin releases in config pre release step review all configs and ensure we use tags digests where appropriate
| 1
|
279,243
| 30,702,482,126
|
IssuesEvent
|
2023-07-27 01:33:55
|
nidhi7598/linux-3.0.35_CVE-2018-13405
|
https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2018-13405
|
closed
|
CVE-2019-11884 (Low) detected in linux-stable-rtv3.8.6 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2019-11884 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hidp/sock.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hidp/sock.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hidp/sock.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The do_hidp_sock_ioctl function in net/bluetooth/hidp/sock.c in the Linux kernel before 5.0.15 allows a local user to obtain potentially sensitive information from kernel stack memory via a HIDPCONNADD command, because a name field may not end with a '\0' character.
<p>Publish Date: 2019-05-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11884>CVE-2019-11884</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: 5.0.15</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-11884 (Low) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2019-11884 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hidp/sock.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hidp/sock.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hidp/sock.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The do_hidp_sock_ioctl function in net/bluetooth/hidp/sock.c in the Linux kernel before 5.0.15 allows a local user to obtain potentially sensitive information from kernel stack memory via a HIDPCONNADD command, because a name field may not end with a '\0' character.
<p>Publish Date: 2019-05-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11884>CVE-2019-11884</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: 5.0.15</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve low detected in linux stable autoclosed cve low severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net bluetooth hidp sock c net bluetooth hidp sock c net bluetooth hidp sock c vulnerability details the do hidp sock ioctl function in net bluetooth hidp sock c in the linux kernel before allows a local user to obtain potentially sensitive information from kernel stack memory via a hidpconnadd command because a name field may not end with a character publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
909
| 3,212,520,457
|
IssuesEvent
|
2015-10-06 15:49:42
|
AdguardTeam/AdguardForWindows
|
https://api.github.com/repos/AdguardTeam/AdguardForWindows
|
opened
|
Speed up Adguard initialization
|
Service
|
There are some bottlenecks we should fix.
***Upgrade check (15 sec):***
```
INFO, AdguardSvc, 3, 04.10.2015 10:23:42.805, Open database path=C:\ProgramData\Adguard\adguard.db version=1
INFO, AdguardSvc, 3, 04.10.2015 10:23:46.456, Database file exists, checking if upgrade is needed.
INFO, AdguardSvc, 3, 04.10.2015 10:24:02.180, Database current version is 1
```
**Registering services (3 sec):**
```
INFO, AdguardSvc, 3, 04.10.2015 10:24:02.305, Preference table exists, doing nothing
INFO, AdguardSvc, 3, 04.10.2015 10:24:05.612, Services registered succesfully
```
**Starting API server (4 sec):**
```
INFO, AdguardSvc, 3, 04.10.2015 10:24:19.294, ApplicationApiServer: Starting API server
INFO, AdguardSvc, 3, 04.10.2015 10:24:23.116, Service with contract Adguard.Ipc.Contract.IApplicationApi published on net.pipe://127.0.0.1/AdguardApiEndpoint
```
|
1.0
|
Speed up Adguard initialization - There are some bottlenecks we should fix.
***Upgrade check (15 sec):***
```
INFO, AdguardSvc, 3, 04.10.2015 10:23:42.805, Open database path=C:\ProgramData\Adguard\adguard.db version=1
INFO, AdguardSvc, 3, 04.10.2015 10:23:46.456, Database file exists, checking if upgrade is needed.
INFO, AdguardSvc, 3, 04.10.2015 10:24:02.180, Database current version is 1
```
**Registering services (3 sec):**
```
INFO, AdguardSvc, 3, 04.10.2015 10:24:02.305, Preference table exists, doing nothing
INFO, AdguardSvc, 3, 04.10.2015 10:24:05.612, Services registered succesfully
```
**Starting API server (4 sec):**
```
INFO, AdguardSvc, 3, 04.10.2015 10:24:19.294, ApplicationApiServer: Starting API server
INFO, AdguardSvc, 3, 04.10.2015 10:24:23.116, Service with contract Adguard.Ipc.Contract.IApplicationApi published on net.pipe://127.0.0.1/AdguardApiEndpoint
```
|
non_process
|
speed up adguard initialization there are some bottlenecks we should fix upgrade check sec info adguardsvc open database path c programdata adguard adguard db version info adguardsvc database file exists checking if upgrade is needed info adguardsvc database current version is registering services sec info adguardsvc preference table exists doing nothing info adguardsvc services registered succesfully starting api server sec info adguardsvc applicationapiserver starting api server info adguardsvc service with contract adguard ipc contract iapplicationapi published on net pipe adguardapiendpoint
| 0
|
133,761
| 5,207,818,884
|
IssuesEvent
|
2017-01-25 01:05:09
|
SnirkImmington/protosnirk
|
https://api.github.com/repos/SnirkImmington/protosnirk
|
reopened
|
Logging
|
enhancement low priority
|
Protosnirk is starting to get to the point where it would be nice to have logs of what's going on.
Right now, I have some debug-print statements that show up in `stdout` during tests - this is helpful because there's no command-line invocation of protosnirk, and the output text only shows up if a test fails. I may continue using this style until the split in #8.
This discussion mostly revolves around using the library vs using a cmd line tool, i.e. `protosnirkc`. The latter is blocked in #8.
Here are my thoughts on logging:
- Completely disabled on release builds - perhaps the macros themselves should be behind `#cfg[debug]` or similar
- Print to `stdout` during tests depending on environment variables (default should be `info`, `trace` should be toggled)
- Print to `stderr` in debug builds depending on environment variables (default should be `warn` or `error`, others toggled)
- Environment variables can trigger logging for individual modules
- Log should include file/line number where invoked
- When printing to `stderr` use colors for trace/info/warn
- `error` logs should be redundant
With these requirements in mind, use either the `log` or `slog` crate.
|
1.0
|
Logging - Protosnirk is starting to get to the point where it would be nice to have logs of what's going on.
Right now, I have some debug-print statements that show up in `stdout` during tests - this is helpful because there's no command-line invocation of protosnirk, and the output text only shows up if a test fails. I may continue using this style until the split in #8.
This discussion mostly revolves around using the library vs using a cmd line tool, i.e. `protosnirkc`. The latter is blocked in #8.
Here are my thoughts on logging:
- Completely disabled on release builds - perhaps the macros themselves should be behind `#cfg[debug]` or similar
- Print to `stdout` during tests depending on environment variables (default should be `info`, `trace` should be toggled)
- Print to `stderr` in debug builds depending on environment variables (default should be `warn` or `error`, others toggled)
- Environment variables can trigger logging for individual modules
- Log should include file/line number where invoked
- When printing to `stderr` use colors for trace/info/warn
- `error` logs should be redundant
With these requirements in mind, use either the `log` or `slog` crate.
|
non_process
|
logging protosnirk is starting to get to the point where it would be nice to have logs of what s going on right now i have some debug print statements that show up in stdout during tests this is helpful because there s no command line invocation of protosnirk and the output text only shows up if a test fails i may continue using this style until the split in this discussion mostly revolves around using the library vs using a cmd line tool i e protosnirkc the latter is blocked in here are my thoughts on logging completely disabled on release builds perhaps the macros themselves should be behind cfg or similar print to stdout during tests depending on environment variables default should be info trace should be toggled print to stderr in debug builds depending on environment variables default should be warn or error others toggled environment variables can trigger logging for individual modules log should include file line number where invoked when printing to stderr use colors for trace info warn error logs should be redundant with these requirements in mind use either the log or slog crate
| 0
|
716,706
| 24,644,860,363
|
IssuesEvent
|
2022-10-17 14:12:47
|
jellywallet/extension
|
https://api.github.com/repos/jellywallet/extension
|
closed
|
Check if user completed DFX KYC
|
Task: Subfeature Priority: High
|
KYC should be done on the DFX website, we only need to check if the user already did KYC. If the KYC wasn't done show the link to the user.
On the staking page:
- Call GET /v1/kyc to get the KYC status
- If the status is not complete call POST /v1/kyc to start the KYC process and get the KYC hash
- Using that has form a link to do the KYC on the DFX website: https://payment.dfx.swiss/kyc?code=[kyc_hash]
- Show that link to the user
[API](https://api.dfx.swiss/swagger/#/kyc). All the routes have to be called with the bearer token set in the header (same as when buying and selling).
|
1.0
|
Check if user completed DFX KYC - KYC should be done on the DFX website, we only need to check if the user already did KYC. If the KYC wasn't done show the link to the user.
On the staking page:
- Call GET /v1/kyc to get the KYC status
- If the status is not complete call POST /v1/kyc to start the KYC process and get the KYC hash
- Using that has form a link to do the KYC on the DFX website: https://payment.dfx.swiss/kyc?code=[kyc_hash]
- Show that link to the user
[API](https://api.dfx.swiss/swagger/#/kyc). All the routes have to be called with the bearer token set in the header (same as when buying and selling).
|
non_process
|
check if user completed dfx kyc kyc should be done on the dfx website we only need to check if the user already did kyc if the kyc wasn t done show the link to the user on the staking page call get kyc to get the kyc status if the status is not complete call post kyc to start the kyc process and get the kyc hash using that has form a link to do the kyc on the dfx website show that link to the user all the routes have to be called with the bearer token set in the header same as when buying and selling
| 0
|
19,472
| 25,780,126,612
|
IssuesEvent
|
2022-12-09 15:14:04
|
DSpace/DSpace
|
https://api.github.com/repos/DSpace/DSpace
|
opened
|
`bin/dspace read` recognizes only launcher commands, not "scripts"
|
bug interface: command-line needs triage tools: processes
|
**Describe the bug**
The `read` command fails to recognize `filter-media` and probably any other command that's been converted to a "script". `o.d.launcher.CommandLauncher` uses `ScriptLauncher.runOneCommand()` but probably should use `ScriptLauncher.handleScript()`.
**To Reproduce**
Steps to reproduce the behavior:
1. Prepare a file X of commands which calls a "script"-ified command. One will be enough.
2. `bin/dspace read X` will claim that no such command(s) exist(s).
**Expected behavior**
All commands should be supported by `read`.
|
1.0
|
`bin/dspace read` recognizes only launcher commands, not "scripts" - **Describe the bug**
The `read` command fails to recognize `filter-media` and probably any other command that's been converted to a "script". `o.d.launcher.CommandLauncher` uses `ScriptLauncher.runOneCommand()` but probably should use `ScriptLauncher.handleScript()`.
**To Reproduce**
Steps to reproduce the behavior:
1. Prepare a file X of commands which calls a "script"-ified command. One will be enough.
2. `bin/dspace read X` will claim that no such command(s) exist(s).
**Expected behavior**
All commands should be supported by `read`.
|
process
|
bin dspace read recognizes only launcher commands not scripts describe the bug the read command fails to recognize filter media and probably any other command that s been converted to a script o d launcher commandlauncher uses scriptlauncher runonecommand but probably should use scriptlauncher handlescript to reproduce steps to reproduce the behavior prepare a file x of commands which calls a script ified command one will be enough bin dspace read x will claim that no such command s exist s expected behavior all commands should be supported by read
| 1
|
18,671
| 24,587,615,925
|
IssuesEvent
|
2022-10-13 21:19:25
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
opened
|
Eliminate the need for a internal & public ESLint package
|
type: enhancement process: dependencies npm: @cypress/eslint-plugin-dev
|
### What would you like?
### What
1. Move [Cypess ESLint Dev](https://github.com/cypress-io/cypress/tree/develop/npm/eslint-plugin-dev) rules into the Cypress root and eliminate the need to maintain an internal & public package.
2. Add scripts to repo root and lint rules to individual packages
3. Update dependencies and stale issues
### Why is this needed?
### Why
There are a few things that can be addressed:
1. There is an **internal** ESLint plugin with rules shared across different entities. However it's [published on npm](https://www.npmjs.com/package/@cypress/eslint-plugin-dev) (3,783 weekly DLs π€) in addition to the **public** [version](https://www.npmjs.com/package/eslint-plugin-cypress) (2,378,591 weekly DLs).
2. The internal version is a dependency and its rules are [extended](https://github.com/cypress-io/eslint-plugin-cypress/blob/master/.eslintrc.json) in the public version with [additional custom rules](https://github.com/cypress-io/eslint-plugin-cypress/tree/master/lib/rules).
4. Some lint rules seem to be run at the root level across the Cypress repo, where adding scripts at the root to run lint rules of each lib could be more optimal.
5. We could move the rules from the internal dev plugin into the rules at the repo root, remove the dev package and move the public version into the monorepo's npm packages to not be maintaining two and minimize confusion about which to use.
Also...
1. Users have begun forking the public version to update & make improvements because the current package has gone stale.
2. ESLint and supporting tooling is very outdated in the dev version making authoring new rules with modern .js in the public for users incompatible.
3. There are dependencies in the public version that [aren't being used](https://github.com/cypress-io/eslint-plugin-cypress/blob/9bcf51e333b14750df3e0148995df172fd17ed60/package.json#L29-L30) and could use general cleanup and maintenance by way of Dependebot or Renovate so the outdated dependency issue is automated away.
### Other
_No response_
|
1.0
|
Eliminate the need for a internal & public ESLint package - ### What would you like?
### What
1. Move [Cypess ESLint Dev](https://github.com/cypress-io/cypress/tree/develop/npm/eslint-plugin-dev) rules into the Cypress root and eliminate the need to maintain an internal & public package.
2. Add scripts to repo root and lint rules to individual packages
3. Update dependencies and stale issues
### Why is this needed?
### Why
There are a few things that can be addressed:
1. There is an **internal** ESLint plugin with rules shared across different entities. However it's [published on npm](https://www.npmjs.com/package/@cypress/eslint-plugin-dev) (3,783 weekly DLs π€) in addition to the **public** [version](https://www.npmjs.com/package/eslint-plugin-cypress) (2,378,591 weekly DLs).
2. The internal version is a dependency and its rules are [extended](https://github.com/cypress-io/eslint-plugin-cypress/blob/master/.eslintrc.json) in the public version with [additional custom rules](https://github.com/cypress-io/eslint-plugin-cypress/tree/master/lib/rules).
4. Some lint rules seem to be run at the root level across the Cypress repo, where adding scripts at the root to run lint rules of each lib could be more optimal.
5. We could move the rules from the internal dev plugin into the rules at the repo root, remove the dev package and move the public version into the monorepo's npm packages to not be maintaining two and minimize confusion about which to use.
Also...
1. Users have begun forking the public version to update & make improvements because the current package has gone stale.
2. ESLint and supporting tooling is very outdated in the dev version making authoring new rules with modern .js in the public for users incompatible.
3. There are dependencies in the public version that [aren't being used](https://github.com/cypress-io/eslint-plugin-cypress/blob/9bcf51e333b14750df3e0148995df172fd17ed60/package.json#L29-L30) and could use general cleanup and maintenance by way of Dependebot or Renovate so the outdated dependency issue is automated away.
### Other
_No response_
|
process
|
eliminate the need for a internal public eslint package what would you like what move rules into the cypress root and eliminate the need to maintain an internal public package add scripts to repo root and lint rules to individual packages update dependencies and stale issues why is this needed why there are a few things that can be addressed there is an internal eslint plugin with rules shared across different entities however it s weekly dls π€ in addition to the public weekly dls the internal version is a dependency and its rules are in the public version with some lint rules seem to be run at the root level across the cypress repo where adding scripts at the root to run lint rules of each lib could be more optimal we could move the rules from the internal dev plugin into the rules at the repo root remove the dev package and move the public version into the monorepo s npm packages to not be maintaining two and minimize confusion about which to use also users have begun forking the public version to update make improvements because the current package has gone stale eslint and supporting tooling is very outdated in the dev version making authoring new rules with modern js in the public for users incompatible there are dependencies in the public version that and could use general cleanup and maintenance by way of dependebot or renovate so the outdated dependency issue is automated away other no response
| 1
|
5,096
| 7,704,617,606
|
IssuesEvent
|
2018-05-21 12:57:46
|
chainside/btcpy
|
https://api.github.com/repos/chainside/btcpy
|
closed
|
Pb of syntax with version 0.5
|
compatibility issue
|
Hi,
I've just installed the latest version (0.5) in a Python 3.3 environment.
Installation returns a few warnings suggesting syntax problems.
```
Installing collected packages: chainside-btcpy, ecdsa
Running setup.py install for chainside-btcpy
D:\tools\python333\python.exe c:\users\admin\appdata\local\temp\tmptmxbw0.py
File "D:\tools\python333\Lib\site-packages\btcpy\setup.py", line 26
return func(*args, **kwargs, strict=strict)
^
SyntaxError: invalid syntax
File "D:\tools\python333\Lib\site-packages\btcpy\structs\script.py", line 774
return [int(m), *pubkeys, int(n)]
^
SyntaxError: can use starred expression only as assignment target
File "D:\tools\python333\Lib\site-packages\tests\integration.py", line 132
self.instance = self.get_script_cls()(*self.get_args(), *args, *scripts)
^
SyntaxError: invalid syntax
removing c:\users\admin\appdata\local\temp\tmptmxbw0.py
D:\tools\python333\lib\distutils\dist.py:257: UserWarning: Unknown distribution option: 'python_requires'
warnings.warn(msg)
```
Problem is confirmed if I try to run a script calling the setup() function.
```
File "D:\tools\python333\lib\site-packages\btcpy\setup.py", line 26
return func(*args, **kwargs, strict=strict)
^
SyntaxError: invalid syntax
```
Does it mean that Python 3.3 isn't supported anymore ?
|
True
|
Pb of syntax with version 0.5 - Hi,
I've just installed the latest version (0.5) in a Python 3.3 environment.
Installation returns a few warnings suggesting syntax problems.
```
Installing collected packages: chainside-btcpy, ecdsa
Running setup.py install for chainside-btcpy
D:\tools\python333\python.exe c:\users\admin\appdata\local\temp\tmptmxbw0.py
File "D:\tools\python333\Lib\site-packages\btcpy\setup.py", line 26
return func(*args, **kwargs, strict=strict)
^
SyntaxError: invalid syntax
File "D:\tools\python333\Lib\site-packages\btcpy\structs\script.py", line 774
return [int(m), *pubkeys, int(n)]
^
SyntaxError: can use starred expression only as assignment target
File "D:\tools\python333\Lib\site-packages\tests\integration.py", line 132
self.instance = self.get_script_cls()(*self.get_args(), *args, *scripts)
^
SyntaxError: invalid syntax
removing c:\users\admin\appdata\local\temp\tmptmxbw0.py
D:\tools\python333\lib\distutils\dist.py:257: UserWarning: Unknown distribution option: 'python_requires'
warnings.warn(msg)
```
Problem is confirmed if I try to run a script calling the setup() function.
```
File "D:\tools\python333\lib\site-packages\btcpy\setup.py", line 26
return func(*args, **kwargs, strict=strict)
^
SyntaxError: invalid syntax
```
Does it mean that Python 3.3 isn't supported anymore ?
|
non_process
|
pb of syntax with version hi i ve just installed the latest version in a python environment installation returns a few warnings suggesting syntax problems installing collected packages chainside btcpy ecdsa running setup py install for chainside btcpy d tools python exe c users admin appdata local temp py file d tools lib site packages btcpy setup py line return func args kwargs strict strict syntaxerror invalid syntax file d tools lib site packages btcpy structs script py line return syntaxerror can use starred expression only as assignment target file d tools lib site packages tests integration py line self instance self get script cls self get args args scripts syntaxerror invalid syntax removing c users admin appdata local temp py d tools lib distutils dist py userwarning unknown distribution option python requires warnings warn msg problem is confirmed if i try to run a script calling the setup function file d tools lib site packages btcpy setup py line return func args kwargs strict strict syntaxerror invalid syntax does it mean that python isn t supported anymore
| 0
|
70,785
| 13,531,615,004
|
IssuesEvent
|
2020-09-15 22:00:46
|
DataBiosphere/azul
|
https://api.github.com/repos/DataBiosphere/azul
|
closed
|
Service responds with 500 error if invalid catalog is specified
|
bug code demoed orange
|
I assume this should be a 4xx error.
Request: https://service.dev.singlecell.gi.ucsc.edu/index/files?catalog=foo
response:
```
Traceback (most recent call last):
File "/var/task/chalice/app.py", line 1112, in _get_view_function_response
response = view_function(**function_args)
File "/var/task/app.py", line 1074, in get_data
return repository_search('files', file_id)
File "/var/task/app.py", line 849, in repository_search
return service.get_data(catalog, entity_type, pagination, filters, item_id, file_url)
File "/var/task/azul/service/repository_service.py", line 86, in get_data
return self._get_items(catalog, entity_type, pagination, filters, file_url_func)
File "/var/task/azul/service/repository_service.py", line 62, in _get_items
response = self._get_data(catalog, entity_type, pagination, filters, file_url_func)
File "/var/task/azul/service/repository_service.py", line 43, in _get_data
response = self.transform_request(catalog=catalog,
File "/var/task/azul/service/elasticsearch_service.py", line 545, in transform_request
service_config = self.service_config(catalog)
File "/var/task/azul/service/elasticsearch_service.py", line 93, in service_config
return self._service_config or self.metadata_plugin(catalog).service_config()
File "/var/task/azul/indexer/document_service.py", line 39, in metadata_plugin
return MetadataPlugin.load(catalog).create()
File "/var/task/azul/plugins/__init__.py", line 87, in load
plugin_package_name = config.plugin_name(catalog, plugin_type_name)
File "/var/task/azul/__init__.py", line 618, in plugin_name
return self.catalogs[catalog_name][plugin_type]
KeyError: 'foo'
```
|
1.0
|
Service responds with 500 error if invalid catalog is specified - I assume this should be a 4xx error.
Request: https://service.dev.singlecell.gi.ucsc.edu/index/files?catalog=foo
response:
```
Traceback (most recent call last):
File "/var/task/chalice/app.py", line 1112, in _get_view_function_response
response = view_function(**function_args)
File "/var/task/app.py", line 1074, in get_data
return repository_search('files', file_id)
File "/var/task/app.py", line 849, in repository_search
return service.get_data(catalog, entity_type, pagination, filters, item_id, file_url)
File "/var/task/azul/service/repository_service.py", line 86, in get_data
return self._get_items(catalog, entity_type, pagination, filters, file_url_func)
File "/var/task/azul/service/repository_service.py", line 62, in _get_items
response = self._get_data(catalog, entity_type, pagination, filters, file_url_func)
File "/var/task/azul/service/repository_service.py", line 43, in _get_data
response = self.transform_request(catalog=catalog,
File "/var/task/azul/service/elasticsearch_service.py", line 545, in transform_request
service_config = self.service_config(catalog)
File "/var/task/azul/service/elasticsearch_service.py", line 93, in service_config
return self._service_config or self.metadata_plugin(catalog).service_config()
File "/var/task/azul/indexer/document_service.py", line 39, in metadata_plugin
return MetadataPlugin.load(catalog).create()
File "/var/task/azul/plugins/__init__.py", line 87, in load
plugin_package_name = config.plugin_name(catalog, plugin_type_name)
File "/var/task/azul/__init__.py", line 618, in plugin_name
return self.catalogs[catalog_name][plugin_type]
KeyError: 'foo'
```
|
non_process
|
service responds with error if invalid catalog is specified i assume this should be a error request response traceback most recent call last file var task chalice app py line in get view function response response view function function args file var task app py line in get data return repository search files file id file var task app py line in repository search return service get data catalog entity type pagination filters item id file url file var task azul service repository service py line in get data return self get items catalog entity type pagination filters file url func file var task azul service repository service py line in get items response self get data catalog entity type pagination filters file url func file var task azul service repository service py line in get data response self transform request catalog catalog file var task azul service elasticsearch service py line in transform request service config self service config catalog file var task azul service elasticsearch service py line in service config return self service config or self metadata plugin catalog service config file var task azul indexer document service py line in metadata plugin return metadataplugin load catalog create file var task azul plugins init py line in load plugin package name config plugin name catalog plugin type name file var task azul init py line in plugin name return self catalogs keyerror foo
| 0
|
495,452
| 14,282,158,581
|
IssuesEvent
|
2020-11-23 09:12:15
|
canonical-web-and-design/charmhub.io
|
https://api.github.com/repos/canonical-web-and-design/charmhub.io
|
closed
|
Wider screen size and integration pop-up window
|
Priority: High
|
Wider screen size hides the second charm on the right in the integration pop-up window

|
1.0
|
Wider screen size and integration pop-up window - Wider screen size hides the second charm on the right in the integration pop-up window

|
non_process
|
wider screen size and integration pop up window wider screen size hides the second charm on the right in the integration pop up window
| 0
|
2,916
| 5,914,189,587
|
IssuesEvent
|
2017-05-22 01:10:40
|
AffiliateWP/AffiliateWP
|
https://api.github.com/repos/AffiliateWP/AffiliateWP
|
closed
|
CSV import for affiliates and referrals
|
batch-processing enhancement Has PR needs testing
|
We should provide a way to import affiliates and referrals via a CSV file.
Related #337
PR: #1891
|
1.0
|
CSV import for affiliates and referrals - We should provide a way to import affiliates and referrals via a CSV file.
Related #337
PR: #1891
|
process
|
csv import for affiliates and referrals we should provide a way to import affiliates and referrals via a csv file related pr
| 1
|
16,458
| 21,336,661,086
|
IssuesEvent
|
2022-04-18 15:20:22
|
rdoddanavar/hpr-sim
|
https://api.github.com/repos/rdoddanavar/hpr-sim
|
closed
|
Archive RASAero Data
|
pre-processing
|
Create plain text data format (preferably in a `*.csv`) to archive processed RASAero data; look at `numpy.savetxt` & `numpy.loadtxt`
|
1.0
|
Archive RASAero Data - Create plain text data format (preferably in a `*.csv`) to archive processed RASAero data; look at `numpy.savetxt` & `numpy.loadtxt`
|
process
|
archive rasaero data create plain text data format preferably in a csv to archive processed rasaero data look at numpy savetxt numpy loadtxt
| 1
|
53,404
| 13,806,599,424
|
IssuesEvent
|
2020-10-11 18:21:45
|
sammiearchie77/binaryoptionslimited
|
https://api.github.com/repos/sammiearchie77/binaryoptionslimited
|
opened
|
WS-2019-0209 (Medium) detected in marked-0.3.6.js
|
security vulnerability
|
## WS-2019-0209 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.6.js</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.6/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.6/marked.js</a></p>
<p>Path to dependency file: binaryoptionslimited/front end/admin/crypto-admin-templates.multipurposethemes.com/sass/dark-multi/pages/forms_editor_markdown.html</p>
<p>Path to vulnerable library: binaryoptionslimited/front end/admin/crypto-admin-templates.multipurposethemes.com/sass/dark-multi/pages/forms_editor_markdown.html</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.3.6.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sammiearchie77/binaryoptionslimited/commit/8e72b732c41253218e9e8d5484272ae7105cb17f">8e72b732c41253218e9e8d5484272ae7105cb17f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 0.7.0 vulnerable to Redos attack by he _label subrule that may significantly degrade parsing performance of malformed input.
<p>Publish Date: 2019-07-04
<p>URL: <a href=https://github.com/markedjs/marked/pull/1515>WS-2019-0209</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1076">https://www.npmjs.com/advisories/1076</a></p>
<p>Release Date: 2019-09-05</p>
<p>Fix Resolution: 0.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0209 (Medium) detected in marked-0.3.6.js - ## WS-2019-0209 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.6.js</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.6/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.6/marked.js</a></p>
<p>Path to dependency file: binaryoptionslimited/front end/admin/crypto-admin-templates.multipurposethemes.com/sass/dark-multi/pages/forms_editor_markdown.html</p>
<p>Path to vulnerable library: binaryoptionslimited/front end/admin/crypto-admin-templates.multipurposethemes.com/sass/dark-multi/pages/forms_editor_markdown.html</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.3.6.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sammiearchie77/binaryoptionslimited/commit/8e72b732c41253218e9e8d5484272ae7105cb17f">8e72b732c41253218e9e8d5484272ae7105cb17f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 0.7.0 vulnerable to Redos attack by he _label subrule that may significantly degrade parsing performance of malformed input.
<p>Publish Date: 2019-07-04
<p>URL: <a href=https://github.com/markedjs/marked/pull/1515>WS-2019-0209</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1076">https://www.npmjs.com/advisories/1076</a></p>
<p>Release Date: 2019-09-05</p>
<p>Fix Resolution: 0.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in marked js ws medium severity vulnerability vulnerable library marked js a markdown parser built for speed library home page a href path to dependency file binaryoptionslimited front end admin crypto admin templates multipurposethemes com sass dark multi pages forms editor markdown html path to vulnerable library binaryoptionslimited front end admin crypto admin templates multipurposethemes com sass dark multi pages forms editor markdown html dependency hierarchy x marked js vulnerable library found in head commit a href found in base branch master vulnerability details marked before vulnerable to redos attack by he label subrule that may significantly degrade parsing performance of malformed input publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
11,279
| 14,077,958,415
|
IssuesEvent
|
2020-11-04 12:51:59
|
tommy-josepovic/smarthome-simulator-team-2
|
https://api.github.com/repos/tommy-josepovic/smarthome-simulator-team-2
|
opened
|
D2 - Diagrams
|
software process
|
Diagrams include:
- Update class diagram with the new classes implemented
- Sequence diagram (1 for each new use case) Remember to write the use cases using essential style and at User-goal level.
- State-machine diagram for the House view
- Activity diagram for Smart home security module
|
1.0
|
D2 - Diagrams - Diagrams include:
- Update class diagram with the new classes implemented
- Sequence diagram (1 for each new use case) Remember to write the use cases using essential style and at User-goal level.
- State-machine diagram for the House view
- Activity diagram for Smart home security module
|
process
|
diagrams diagrams include update class diagram with the new classes implemented sequence diagram for each new use case remember to write the use cases using essential style and at user goal level state machine diagram for the house view activity diagram for smart home security module
| 1
|
6,964
| 10,118,698,708
|
IssuesEvent
|
2019-07-31 09:39:18
|
BlesseNtumble/GalaxySpace
|
https://api.github.com/repos/BlesseNtumble/GalaxySpace
|
closed
|
ΠΡΠ³ΠΎΠ²Π°Ρ ΠΏΠ΅ΡΡ Π½Π΅ ΡΠ΄Π²Π°ΠΈΠ²Π°Π΅Ρ Π²ΡΡ
ΠΎΠ΄ ΡΠ»ΠΈΡΠΊΠΎΠ² Ρ Π½Π΅ΠΊΠΎΡΠΎΡΡΡ
ΡΡΠ΄
|
1.12.2 in the process of correcting
|
**Please fill in the form below:**
------------------------------------------------------------------------
1. Minecraft version: 1.12.2
2. Galacticraft version: 4.0.2.211
3. GalaxySpace version: 2.0.8
4. AsmodeusCore version (for 2.0.1 version and above): 0.0.8
5. Side (Single player (SSP), Multiplayer (SMP), or SSP opened to LAN (LAN)): SSP
Description of the issue:
------------------------------------------------------------------------
Π‘ΠΎΠ³Π»Π°ΡΠ½ΠΎ Galacticraft Wiki, [ΠΠ»Π΅ΠΊΡΡΠΈΡΠ΅ΡΠΊΠ°Ρ Π΄ΡΠ³ΠΎΠ²Π°Ρ ΠΏΠ΅ΡΡ](https://wiki.micdoodle8.com/wiki/Electric_Arc_Furnace) ΡΠ΄Π²Π°ΠΈΠ²Π°Π΅Ρ Π²ΡΡ
ΠΎΠ΄ ΡΠ»ΠΈΡΠΊΠΎΠ² Ρ ΡΡΠ΄. ΠΠΎ Ρ ΠΠΎΠ±Π°Π»ΡΡΠΎΠ²ΠΎΠΉ, ΠΠΈΠΊΠ΅Π»Π΅Π²ΠΎΠΉ ΠΈ ΠΠ°Π³Π½ΠΈΠ΅Π²ΠΎΠΉ ΡΡΠ΄Π°ΠΌΠΈ ΡΡΠΎ Π½Π΅ ΡΠ°Π±ΠΎΡΠ°Π΅Ρ
Screenshot/Video:
------------------------------------------------------------------------
N/A
Attached log file (or url on pastebin.com):
------------------------------------------------------------------------
N/A
|
1.0
|
ΠΡΠ³ΠΎΠ²Π°Ρ ΠΏΠ΅ΡΡ Π½Π΅ ΡΠ΄Π²Π°ΠΈΠ²Π°Π΅Ρ Π²ΡΡ
ΠΎΠ΄ ΡΠ»ΠΈΡΠΊΠΎΠ² Ρ Π½Π΅ΠΊΠΎΡΠΎΡΡΡ
ΡΡΠ΄ - **Please fill in the form below:**
------------------------------------------------------------------------
1. Minecraft version: 1.12.2
2. Galacticraft version: 4.0.2.211
3. GalaxySpace version: 2.0.8
4. AsmodeusCore version (for 2.0.1 version and above): 0.0.8
5. Side (Single player (SSP), Multiplayer (SMP), or SSP opened to LAN (LAN)): SSP
Description of the issue:
------------------------------------------------------------------------
Π‘ΠΎΠ³Π»Π°ΡΠ½ΠΎ Galacticraft Wiki, [ΠΠ»Π΅ΠΊΡΡΠΈΡΠ΅ΡΠΊΠ°Ρ Π΄ΡΠ³ΠΎΠ²Π°Ρ ΠΏΠ΅ΡΡ](https://wiki.micdoodle8.com/wiki/Electric_Arc_Furnace) ΡΠ΄Π²Π°ΠΈΠ²Π°Π΅Ρ Π²ΡΡ
ΠΎΠ΄ ΡΠ»ΠΈΡΠΊΠΎΠ² Ρ ΡΡΠ΄. ΠΠΎ Ρ ΠΠΎΠ±Π°Π»ΡΡΠΎΠ²ΠΎΠΉ, ΠΠΈΠΊΠ΅Π»Π΅Π²ΠΎΠΉ ΠΈ ΠΠ°Π³Π½ΠΈΠ΅Π²ΠΎΠΉ ΡΡΠ΄Π°ΠΌΠΈ ΡΡΠΎ Π½Π΅ ΡΠ°Π±ΠΎΡΠ°Π΅Ρ
Screenshot/Video:
------------------------------------------------------------------------
N/A
Attached log file (or url on pastebin.com):
------------------------------------------------------------------------
N/A
|
process
|
Π΄ΡΠ³ΠΎΠ²Π°Ρ ΠΏΠ΅ΡΡ Π½Π΅ ΡΠ΄Π²Π°ΠΈΠ²Π°Π΅Ρ Π²ΡΡ
ΠΎΠ΄ ΡΠ»ΠΈΡΠΊΠΎΠ² Ρ Π½Π΅ΠΊΠΎΡΠΎΡΡΡ
ΡΡΠ΄ please fill in the form below minecraft version galacticraft version galaxyspace version asmodeuscore version for version and above side single player ssp multiplayer smp or ssp opened to lan lan ssp description of the issue ΡΠΎΠ³Π»Π°ΡΠ½ΠΎ galacticraft wiki ΡΠ΄Π²Π°ΠΈΠ²Π°Π΅Ρ Π²ΡΡ
ΠΎΠ΄ ΡΠ»ΠΈΡΠΊΠΎΠ² Ρ ΡΡΠ΄ Π½ΠΎ Ρ ΠΊΠΎΠ±Π°Π»ΡΡΠΎΠ²ΠΎΠΉ Π½ΠΈΠΊΠ΅Π»Π΅Π²ΠΎΠΉ ΠΈ ΠΌΠ°Π³Π½ΠΈΠ΅Π²ΠΎΠΉ ΡΡΠ΄Π°ΠΌΠΈ ΡΡΠΎ Π½Π΅ ΡΠ°Π±ΠΎΡΠ°Π΅Ρ screenshot video n a attached log file or url on pastebin com n a
| 1
|
305,976
| 9,379,147,260
|
IssuesEvent
|
2019-04-04 14:22:24
|
forpdi/forpdi
|
https://api.github.com/repos/forpdi/forpdi
|
closed
|
Incluir mensagem de confirmaΓ§Γ£o na exclusΓ£o de aΓ§Γ£o de prevenΓ§Γ£o
|
ForRisco enhancement mediumpriority
|
- [x] No risco, sugiro a criaΓ§Γ£o de uma mensagem de confirmaΓ§Γ£o para a exclusΓ£o de uma aΓ§Γ£o de prevenΓ§Γ£o. Atualmente a exclusΓ£o Γ© feita direto.
A mensagem pode ser: "Deseja realmente excluir a aΓ§Γ£o de prevenΓ§Γ£o?" com botΓ΅es de "Sim" e "NΓ£o". Se o usuΓ‘rio clicar em sim a aΓ§Γ£o Γ© excluΓda, se clicar em nΓ£o a exclusΓ£o Γ© cancelada.

- [x] O mesmo ocorre para monitoramento. Desenvolver a mesma melhoria.
- [x] O mesmo ocorre para os incidentes. Desenvolver a mesma melhoria.
- [x] O mesmo ocorre para os contingenciamentos. Desenvolver a mesma melhoria.
|
1.0
|
Incluir mensagem de confirmaΓ§Γ£o na exclusΓ£o de aΓ§Γ£o de prevenΓ§Γ£o - - [x] No risco, sugiro a criaΓ§Γ£o de uma mensagem de confirmaΓ§Γ£o para a exclusΓ£o de uma aΓ§Γ£o de prevenΓ§Γ£o. Atualmente a exclusΓ£o Γ© feita direto.
A mensagem pode ser: "Deseja realmente excluir a aΓ§Γ£o de prevenΓ§Γ£o?" com botΓ΅es de "Sim" e "NΓ£o". Se o usuΓ‘rio clicar em sim a aΓ§Γ£o Γ© excluΓda, se clicar em nΓ£o a exclusΓ£o Γ© cancelada.

- [x] O mesmo ocorre para monitoramento. Desenvolver a mesma melhoria.
- [x] O mesmo ocorre para os incidentes. Desenvolver a mesma melhoria.
- [x] O mesmo ocorre para os contingenciamentos. Desenvolver a mesma melhoria.
|
non_process
|
incluir mensagem de confirmaΓ§Γ£o na exclusΓ£o de aΓ§Γ£o de prevenΓ§Γ£o no risco sugiro a criaΓ§Γ£o de uma mensagem de confirmaΓ§Γ£o para a exclusΓ£o de uma aΓ§Γ£o de prevenΓ§Γ£o atualmente a exclusΓ£o Γ© feita direto a mensagem pode ser deseja realmente excluir a aΓ§Γ£o de prevenΓ§Γ£o com botΓ΅es de sim e nΓ£o se o usuΓ‘rio clicar em sim a aΓ§Γ£o Γ© excluΓda se clicar em nΓ£o a exclusΓ£o Γ© cancelada o mesmo ocorre para monitoramento desenvolver a mesma melhoria o mesmo ocorre para os incidentes desenvolver a mesma melhoria o mesmo ocorre para os contingenciamentos desenvolver a mesma melhoria
| 0
|
2,020
| 4,846,103,162
|
IssuesEvent
|
2016-11-10 10:36:17
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Spawning child processes in bash on Windows can crash the parent process
|
child_process windows
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v7.1.0
* **Platform**: Linux OWNER43-PC 3.4.0+ #1 PREEMPT Thu Aug 1 17:06:05 CST 2013 x86_64 x86_64 x86_64 GNU/Linux, Windows 10 Pro
* **Subsystem**: child_process
<!-- Enter your issue details below this comment. -->
I wrote a short piece of code to spawn a child process:
```javascript
const {spawn} = require('child_process');
const sockets = spawn('go', ['run', 'sockets.go']);
```
That resulted in this error:
```
CRASH: Error: err >= 0
at Object.exports._errnoException (util.js:1018:20)
at exports._exceptionWithHostPort (util.js:1045:20)
at PipeConnectWrap.afterConnect [as oncomplete] (net.js:1090:14)
```
I tested the same piece of code on a FreeBSD 11.0 system using the same versions of bash, node, and go, which resulted in no crash, so I'm fairly sure the issue's in bash on Windows. Would there be a way to possibly mitigate this?
|
1.0
|
Spawning child processes in bash on Windows can crash the parent process - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v7.1.0
* **Platform**: Linux OWNER43-PC 3.4.0+ #1 PREEMPT Thu Aug 1 17:06:05 CST 2013 x86_64 x86_64 x86_64 GNU/Linux, Windows 10 Pro
* **Subsystem**: child_process
<!-- Enter your issue details below this comment. -->
I wrote a short piece of code to spawn a child process:
```javascript
const {spawn} = require('child_process');
const sockets = spawn('go', ['run', 'sockets.go']);
```
That resulted in this error:
```
CRASH: Error: err >= 0
at Object.exports._errnoException (util.js:1018:20)
at exports._exceptionWithHostPort (util.js:1045:20)
at PipeConnectWrap.afterConnect [as oncomplete] (net.js:1090:14)
```
I tested the same piece of code on a FreeBSD 11.0 system using the same versions of bash, node, and go, which resulted in no crash, so I'm fairly sure the issue's in bash on Windows. Would there be a way to possibly mitigate this?
|
process
|
spawning child processes in bash on windows can crash the parent process thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version platform linux pc preempt thu aug cst gnu linux windows pro subsystem child process i wrote a short piece of code to spawn a child process javascript const spawn require child process const sockets spawn go that resulted in this error crash error err at object exports errnoexception util js at exports exceptionwithhostport util js at pipeconnectwrap afterconnect net js i tested the same piece of code on a freebsd system using the same versions of bash node and go which resulted in no crash so i m fairly sure the issue s in bash on windows would there be a way to possibly mitigate this
| 1
|
1,841
| 4,646,980,248
|
IssuesEvent
|
2016-10-01 07:02:46
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Investigate flaky parallel/test-tick-processor-unknown
|
process test tools
|
* **Version**: master
* **Platform**: smartos, windows
* **Subsystem**: process, tools
I've recently started seeing `test-tick-processor-unknown` failures on `smartos14-32` and various Windows configurations in CI.
Most are merely timeouts, but I did see this instance on `smartos14-32` tonight that resulted in a [different result](https://ci.nodejs.org/job/node-test-commit-smartos/4551/nodes=smartos14-32/console):
```
not ok 1174 parallel/test-tick-processor-unknown
# TIMEOUT
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: 8f60a23e
# 10: a8c55c7b
# 11: a8c563ae
# 12: a8c3c37c
# 13: a8c3c084
# 14: a8c3bf5f
# 15: a8c4920b
# 16: a8c36f64
# 17: a8c36a4e
# 18: a8c3686e
# 19: a8c18962
# 20: a8c18a5f
# 21: 8f60b6b6
# 22: a8c149cf
# 23: 8f60b6b6
# 24: 8f66537d
# 25: 8f664baf
# 26: 8f663e0e
# 27: 8f66198a
# 28: 8f63e83e
# 29: 8f627878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: a9c0a23e
# 10: bb85613b
# 11: bb85686e
# 12: bb83c37c
# 13: bb83c084
# 14: bb83bf5f
# 15: bb8457ab
# 16: bb836f64
# 17: bb836a4e
# 18: bb83686e
# 19: bb818962
# 20: bb818a5f
# 21: a9c0b6b6
# 22: bb8149cf
# 23: a9c0b6b6
# 24: a9c6537d
# 25: a9c64baf
# 26: a9c63e0e
# 27: a9c6198a
# 28: a9c3e83e
# 29: a9c27878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: a2d0a23e
# 10: 94c57adb
# 11: 94c5820e
# 12: 94c3c37c
# 13: 94c3c084
# 14: 94c3bf5f
# 15: 94c4832b
# 16: 94c36f64
# 17: 94c36a4e
# 18: 94c3686e
# 19: 94c18962
# 20: 94c18a5f
# 21: a2d0b6b6
# 22: 94c149cf
# 23: a2d0b6b6
# 24: a2d6537d
# 25: a2d64baf
# 26: a2d63e0e
# 27: a2d6198a
# 28: a2d3e83e
# 29: a2d27878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: 8260a23e
# 10: 951560bb
# 11: 951567f9
# 12: 9513c37c
# 13: 9513c084
# 14: 9513bf5f
# 15: 9514a5cb
# 16: 95136f64
# 17: 95136a4e
# 18: 9513686e
# 19: 95118962
# 20: 95118a5f
# 21: 8260b6b6
# 22: 951149cf
# 23: 8260b6b6
# 24: 8266537d
# 25: 82664baf
# 26: 82663e0e
# 27: 8266198a
# 28: 8263e83e
# 29: 82627878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
---
duration_ms: 60.105
```
|
1.0
|
Investigate flaky parallel/test-tick-processor-unknown - * **Version**: master
* **Platform**: smartos, windows
* **Subsystem**: process, tools
I've recently started seeing `test-tick-processor-unknown` failures on `smartos14-32` and various Windows configurations in CI.
Most are merely timeouts, but I did see this instance on `smartos14-32` tonight that resulted in a [different result](https://ci.nodejs.org/job/node-test-commit-smartos/4551/nodes=smartos14-32/console):
```
not ok 1174 parallel/test-tick-processor-unknown
# TIMEOUT
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: 8f60a23e
# 10: a8c55c7b
# 11: a8c563ae
# 12: a8c3c37c
# 13: a8c3c084
# 14: a8c3bf5f
# 15: a8c4920b
# 16: a8c36f64
# 17: a8c36a4e
# 18: a8c3686e
# 19: a8c18962
# 20: a8c18a5f
# 21: 8f60b6b6
# 22: a8c149cf
# 23: 8f60b6b6
# 24: 8f66537d
# 25: 8f664baf
# 26: 8f663e0e
# 27: 8f66198a
# 28: 8f63e83e
# 29: 8f627878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: a9c0a23e
# 10: bb85613b
# 11: bb85686e
# 12: bb83c37c
# 13: bb83c084
# 14: bb83bf5f
# 15: bb8457ab
# 16: bb836f64
# 17: bb836a4e
# 18: bb83686e
# 19: bb818962
# 20: bb818a5f
# 21: a9c0b6b6
# 22: bb8149cf
# 23: a9c0b6b6
# 24: a9c6537d
# 25: a9c64baf
# 26: a9c63e0e
# 27: a9c6198a
# 28: a9c3e83e
# 29: a9c27878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: a2d0a23e
# 10: 94c57adb
# 11: 94c5820e
# 12: 94c3c37c
# 13: 94c3c084
# 14: 94c3bf5f
# 15: 94c4832b
# 16: 94c36f64
# 17: 94c36a4e
# 18: 94c3686e
# 19: 94c18962
# 20: 94c18a5f
# 21: a2d0b6b6
# 22: 94c149cf
# 23: a2d0b6b6
# 24: a2d6537d
# 25: a2d64baf
# 26: a2d63e0e
# 27: a2d6198a
# 28: a2d3e83e
# 29: a2d27878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# FATAL ERROR: invalid array length Allocation failed - JavaScript heap out of memory
# 1: node::Abort() [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 2: node::OnFatalError(char const*, char const*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 3: v8::Utils::ReportOOMFailure(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 5: v8::internal::Heap::AllocateUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 6: v8::internal::Factory::NewUninitializedFixedArray(int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 7: v8::internal::(anonymous namespace)::ElementsAccessorBase<v8::internal::(anonymous namespace)::FastPackedSmiElementsAccessor, v8::internal::(anonymous namespace)::ElementsKindTraits<(v8::internal::ElementsKind)0> >::GrowCapacityAndConvertImpl(v8::internal::Handle<v8::internal::JSObject>, unsigned int) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 8: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 9: 8260a23e
# 10: 951560bb
# 11: 951567f9
# 12: 9513c37c
# 13: 9513c084
# 14: 9513bf5f
# 15: 9514a5cb
# 16: 95136f64
# 17: 95136a4e
# 18: 9513686e
# 19: 95118962
# 20: 95118a5f
# 21: 8260b6b6
# 22: 951149cf
# 23: 8260b6b6
# 24: 8266537d
# 25: 82664baf
# 26: 82663e0e
# 27: 8266198a
# 28: 8263e83e
# 29: 82627878
# 30: v8::internal::Execution::Call(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, int, v8::internal::Handle<v8::internal::Object>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 31: v8::Function::Call(v8::Local<v8::Context>, v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 32: v8::Function::Call(v8::Local<v8::Value>, int, v8::Local<v8::Value>*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 33: node::LoadEnvironment(node::Environment*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 34: node::StartNodeInstance(void*) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 35: node::Start(int, char**) [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 36: main [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
# 37: _start [/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos14-32/out/Release/node]
---
duration_ms: 60.105
```
|
process
|
investigate flaky parallel test tick processor unknown version master platform smartos windows subsystem process tools i ve recently started seeing test tick processor unknown failures on and various windows configurations in ci most are merely timeouts but i did see this instance on tonight that resulted in a not ok parallel test tick processor unknown timeout fatal error invalid array length allocation failed javascript heap out of memory node abort node onfatalerror char const char const utils reportoomfailure char const bool internal fatalprocessoutofmemory char const bool internal heap allocateuninitializedfixedarray int internal factory newuninitializedfixedarray int internal anonymous namespace elementsaccessorbase growcapacityandconvertimpl internal handle unsigned int internal runtime growarrayelements int internal object internal isolate internal execution call internal isolate internal handle internal handle int internal handle function call local local int local function call local int local node loadenvironment node environment node startnodeinstance void node start int char main start fatal error invalid array length allocation failed javascript heap out of memory node abort node onfatalerror char const char const utils reportoomfailure char const bool internal fatalprocessoutofmemory char const bool internal heap allocateuninitializedfixedarray int internal factory newuninitializedfixedarray int internal anonymous namespace elementsaccessorbase growcapacityandconvertimpl internal handle unsigned int internal runtime growarrayelements int internal object internal isolate internal execution call internal isolate internal handle internal handle int internal handle function call local local int local function call local int local node loadenvironment node environment node startnodeinstance void node start int char main start fatal error invalid array length allocation failed javascript heap out of memory node abort node onfatalerror char const char const utils reportoomfailure char const bool internal fatalprocessoutofmemory char const bool internal heap allocateuninitializedfixedarray int internal factory newuninitializedfixedarray int internal anonymous namespace elementsaccessorbase growcapacityandconvertimpl internal handle unsigned int internal runtime growarrayelements int internal object internal isolate internal execution call internal isolate internal handle internal handle int internal handle function call local local int local function call local int local node loadenvironment node environment node startnodeinstance void node start int char main start fatal error invalid array length allocation failed javascript heap out of memory node abort node onfatalerror char const char const utils reportoomfailure char const bool internal fatalprocessoutofmemory char const bool internal heap allocateuninitializedfixedarray int internal factory newuninitializedfixedarray int internal anonymous namespace elementsaccessorbase growcapacityandconvertimpl internal handle unsigned int internal runtime growarrayelements int internal object internal isolate internal execution call internal isolate internal handle internal handle int internal handle function call local local int local function call local int local node loadenvironment node environment node startnodeinstance void node start int char main start duration ms
| 1
|
35,384
| 7,724,264,723
|
IssuesEvent
|
2018-05-24 14:40:03
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
API Cryptokeys DELETE key behaviour
|
auth defect
|
- Program: Authoritative
- Issue type: Bug report
### Short description
When executing the api call to delete a cryptokey, it returns 200 OK instead of 204 No Content on deleting an existing key. But Also returns 200 OK on non-existing keys.
### Environment
- Operating system: CentOS Linux release 7.4.1708 (Core)
- Software version: PowerDNS Auth 4.1.2
- Software source: PowerDNS repository
- Backend source: mysql Ver 15.1 Distrib 5.5.56-MariaDB, for Linux (x86_64) using readline 5.1
### Steps to reproduce
1. Create and sign a domain, in my example dennis6.nl
2. Retrieve all cryptokeys of an zone using the api call "GET /servers/{server_id}/zones/{zone_id}/cryptokeys"
3. Delete an existing key of the zone, using the following api call:
"DELETE /servers/{server_id}/zones/{zone_id}/cryptokeys/{cryptokey_id}"
4. Delete an non-existing key of the zone, using the following api call: "DELETE /servers/{server_id}/zones/{zone_id}/cryptokeys/{cryptokey_id}"
### Expected behaviour
Retrieve all cryptokeys for dennis6.nl:
```
$ php cryptokeys_zone.php dennis6.nl
0 -> id = 133
1 -> id = 145
```
First we delete an existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 145
StatusCode is: 204 No Content
```
Second we delete an non-existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 1454211
StatusCode is: 404 Not Found
```
Or something like "Key doesn't exists", but deleting an non-existing zone also gives 404 Not Found.
### Actual behaviour
Retrieve all cryptokeys for dennis6.nl:
```
$ php cryptokeys_zone.php dennis6.nl
0 -> id = 133
1 -> id = 145
```
3. First we delete an existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 145
StatusCode is: 200 OK
```
> HTTP: Result for "/api/v1/servers/localhost/zones/dennis6.nl/cryptokeys/145": 200, body length: 0
This should be 204 as stated in the documentation:
4. Second we delete an non-existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 1454211
StatusCode is: 200 OK
```
>HTTP: Result for "/api/v1/servers/localhost/zones/dennis6.nl/cryptokeys/1454211": 200, body length: 0
|
1.0
|
API Cryptokeys DELETE key behaviour - - Program: Authoritative
- Issue type: Bug report
### Short description
When executing the api call to delete a cryptokey, it returns 200 OK instead of 204 No Content on deleting an existing key. But Also returns 200 OK on non-existing keys.
### Environment
- Operating system: CentOS Linux release 7.4.1708 (Core)
- Software version: PowerDNS Auth 4.1.2
- Software source: PowerDNS repository
- Backend source: mysql Ver 15.1 Distrib 5.5.56-MariaDB, for Linux (x86_64) using readline 5.1
### Steps to reproduce
1. Create and sign a domain, in my example dennis6.nl
2. Retrieve all cryptokeys of an zone using the api call "GET /servers/{server_id}/zones/{zone_id}/cryptokeys"
3. Delete an existing key of the zone, using the following api call:
"DELETE /servers/{server_id}/zones/{zone_id}/cryptokeys/{cryptokey_id}"
4. Delete an non-existing key of the zone, using the following api call: "DELETE /servers/{server_id}/zones/{zone_id}/cryptokeys/{cryptokey_id}"
### Expected behaviour
Retrieve all cryptokeys for dennis6.nl:
```
$ php cryptokeys_zone.php dennis6.nl
0 -> id = 133
1 -> id = 145
```
First we delete an existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 145
StatusCode is: 204 No Content
```
Second we delete an non-existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 1454211
StatusCode is: 404 Not Found
```
Or something like "Key doesn't exists", but deleting an non-existing zone also gives 404 Not Found.
### Actual behaviour
Retrieve all cryptokeys for dennis6.nl:
```
$ php cryptokeys_zone.php dennis6.nl
0 -> id = 133
1 -> id = 145
```
3. First we delete an existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 145
StatusCode is: 200 OK
```
> HTTP: Result for "/api/v1/servers/localhost/zones/dennis6.nl/cryptokeys/145": 200, body length: 0
This should be 204 as stated in the documentation:
4. Second we delete an non-existing key:
```
$ php delete_cryptokey_zone.php dennis6.nl 1454211
StatusCode is: 200 OK
```
>HTTP: Result for "/api/v1/servers/localhost/zones/dennis6.nl/cryptokeys/1454211": 200, body length: 0
|
non_process
|
api cryptokeys delete key behaviour program authoritative issue type bug report short description when executing the api call to delete a cryptokey it returns ok instead of no content on deleting an existing key but also returns ok on non existing keys environment operating system centos linux release core software version powerdns auth software source powerdns repository backend source mysql ver distrib mariadb for linux using readline steps to reproduce create and sign a domain in my example nl retrieve all cryptokeys of an zone using the api call get servers server id zones zone id cryptokeys delete an existing key of the zone using the following api call delete servers server id zones zone id cryptokeys cryptokey id delete an non existing key of the zone using the following api call delete servers server id zones zone id cryptokeys cryptokey id expected behaviour retrieve all cryptokeys for nl php cryptokeys zone php nl id id first we delete an existing key php delete cryptokey zone php nl statuscode is no content second we delete an non existing key php delete cryptokey zone php nl statuscode is not found or something like key doesn t exists but deleting an non existing zone also gives not found actual behaviour retrieve all cryptokeys for nl php cryptokeys zone php nl id id first we delete an existing key php delete cryptokey zone php nl statuscode is ok http result for api servers localhost zones nl cryptokeys body length this should be as stated in the documentation second we delete an non existing key php delete cryptokey zone php nl statuscode is ok http result for api servers localhost zones nl cryptokeys body length
| 0
|
183,637
| 14,948,078,604
|
IssuesEvent
|
2021-01-26 09:35:08
|
handsontable/handsontable
|
https://api.github.com/repos/handsontable/handsontable
|
closed
|
Inconsistent behaviour of AutoColumnSize and AutoRowSize
|
Auto column size Auto row size Type: Documentation needed Type: Improvement suggestion scroll
|
### Description
For unknown to me reasons, we use a completely different way to define when and how `AutoColumnSize` and `AutoRowSize` should work. Because these plugins are similar it would be better to keep its configuration similar too.
Look at default settings in the following code:
https://github.com/handsontable/handsontable/blob/b9e14ee9c45dc8f42e23414c93e82e6b805f6879/src/defaultSettings.js#L2197-L2248
Particularly at:
```
(...) Default value is `undefined`, which has the same effect as `true` (...)
...
@default {syncLimit: 50}
```
for `AutoColumnSize`
and
```
(...) Default value is `undefined`, which has the same effect as `false` (disabled) (...)
...
@default {syncLimit: 500}
```
for `AutoRowSize`.
Moreover, `GhostTable` has the option to use headers in calculations but you cannot use it to calculate multiline `rowHeaders` in `AutoRowSize`.
Related with: #4122
|
1.0
|
Inconsistent behaviour of AutoColumnSize and AutoRowSize - ### Description
For unknown to me reasons, we use a completely different way to define when and how `AutoColumnSize` and `AutoRowSize` should work. Because these plugins are similar it would be better to keep its configuration similar too.
Look at default settings in the following code:
https://github.com/handsontable/handsontable/blob/b9e14ee9c45dc8f42e23414c93e82e6b805f6879/src/defaultSettings.js#L2197-L2248
Particularly at:
```
(...) Default value is `undefined`, which has the same effect as `true` (...)
...
@default {syncLimit: 50}
```
for `AutoColumnSize`
and
```
(...) Default value is `undefined`, which has the same effect as `false` (disabled) (...)
...
@default {syncLimit: 500}
```
for `AutoRowSize`.
Moreover, `GhostTable` has the option to use headers in calculations but you cannot use it to calculate multiline `rowHeaders` in `AutoRowSize`.
Related with: #4122
|
non_process
|
inconsistent behaviour of autocolumnsize and autorowsize description for unknown to me reasons we use a completely different way to define when and how autocolumnsize and autorowsize should work because these plugins are similar it would be better to keep its configuration similar too look at default settings in the following code particularly at default value is undefined which has the same effect as true default synclimit for autocolumnsize and default value is undefined which has the same effect as false disabled default synclimit for autorowsize moreover ghosttable has the option to use headers in calculations but you cannot use it to calculate multiline rowheaders in autorowsize related with
| 0
|
358,883
| 25,207,293,146
|
IssuesEvent
|
2022-11-13 20:48:16
|
SOunit/RPG-Project
|
https://api.github.com/repos/SOunit/RPG-Project
|
closed
|
require
|
documentation
|
- require some component
```
namespace RPG.Combat
{
[RequireComponent(typeof (Health))]
public class CombatTarget : MonoBehaviour { }
}
```
|
1.0
|
require - - require some component
```
namespace RPG.Combat
{
[RequireComponent(typeof (Health))]
public class CombatTarget : MonoBehaviour { }
}
```
|
non_process
|
require require some component namespace rpg combat public class combattarget monobehaviour
| 0
|
18,640
| 24,580,791,569
|
IssuesEvent
|
2022-10-13 15:27:55
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[FHIR store] Questionnaire response > Text choice > Single select > Answer option is not getting displayed when participant selects 'Other' answer option
|
Bug P0 Response datastore Process: Fixed Process: Tested QA Process: Tested dev
|
Steps:
1. SB> Add/edit questionnaire > Add text choice questionnaire with 'Other' option and single select
2. Launch or publish the updates
3. Sign up or sign in to the mobile app
4. Enroll to the study
5. Submit the response by selecting the 'Other' answer option
6. Go to the FHIR store and observe
AR: 'Other' answer option is not getting stored in the FHIR store
ER: 'Other' answer option should get stored in the FHIR store

|
3.0
|
[FHIR store] Questionnaire response > Text choice > Single select > Answer option is not getting displayed when participant selects 'Other' answer option - Steps:
1. SB> Add/edit questionnaire > Add text choice questionnaire with 'Other' option and single select
2. Launch or publish the updates
3. Sign up or sign in to the mobile app
4. Enroll to the study
5. Submit the response by selecting the 'Other' answer option
6. Go to the FHIR store and observe
AR: 'Other' answer option is not getting stored in the FHIR store
ER: 'Other' answer option should get stored in the FHIR store

|
process
|
questionnaire response text choice single select answer option is not getting displayed when participant selects other answer option steps sb add edit questionnaire add text choice questionnaire with other option and single select launch or publish the updates sign up or sign in to the mobile app enroll to the study submit the response by selecting the other answer option go to the fhir store and observe ar other answer option is not getting stored in the fhir store er other answer option should get stored in the fhir store
| 1
|
493,809
| 14,238,601,760
|
IssuesEvent
|
2020-11-18 18:52:26
|
dtcenter/METplus
|
https://api.github.com/repos/dtcenter/METplus
|
closed
|
Add plot_data_plane wrapper
|
alert: NEED MORE DEFINITION component: use case wrapper priority: medium requestor: NCAR type: new feature
|
Implement a wrapper for plot_data_plane.
## Describe the New Feature ##
*Provide a description of the new feature request here.*
### Acceptance Testing ###
*List input data types and sources.*
*Describe tests required for new functionality.*
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the new feature down into sub-issues.
- [x] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required
- [x] Select **scientist(s)** or **no scientist** required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [x] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## New Feature Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above, including the **Time Estimate** and **Funding source**.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [x] Complete the development and test your changes.
- [x] Add/update log messages for easier debugging.
- [x] Add/update unit tests.
- [x] Add/update documentation.
- [x] Push local changes to GitHub.
- [x] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [x] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [x] Iterate until the reviewer(s) accept and merge your changes.
- [x] Delete your fork or branch.
- [x] Close this issue.
|
1.0
|
Add plot_data_plane wrapper - Implement a wrapper for plot_data_plane.
## Describe the New Feature ##
*Provide a description of the new feature request here.*
### Acceptance Testing ###
*List input data types and sources.*
*Describe tests required for new functionality.*
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the new feature down into sub-issues.
- [x] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required
- [x] Select **scientist(s)** or **no scientist** required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [x] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## New Feature Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above, including the **Time Estimate** and **Funding source**.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [x] Complete the development and test your changes.
- [x] Add/update log messages for easier debugging.
- [x] Add/update unit tests.
- [x] Add/update documentation.
- [x] Push local changes to GitHub.
- [x] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [x] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [x] Iterate until the reviewer(s) accept and merge your changes.
- [x] Delete your fork or branch.
- [x] Close this issue.
|
non_process
|
add plot data plane wrapper implement a wrapper for plot data plane describe the new feature provide a description of the new feature request here acceptance testing list input data types and sources describe tests required for new functionality time estimate estimate the amount of work required here issues should represent approximately to days of work sub issues consider breaking the new feature down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone review projects and select relevant repository and organization ones or add alert need project assignment label select milestone to next major version milestone or future versions define related issue s consider the impact to the other metplus components new feature checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s project s milestone and linked issues iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
| 0
|
21,293
| 28,489,511,897
|
IssuesEvent
|
2023-04-18 10:13:02
|
aiidateam/aiida-core
|
https://api.github.com/repos/aiidateam/aiida-core
|
opened
|
Add the `ProcessNode.exit_code` property
|
priority/nice-to-have topic/orm topic/processes
|
This property should automatically reconstitute an `ExitCode` instance from the `exit_status` and `exit_message` attributes, if defined. This is very useful in `Parser` implementations that need to return the `ExitCode` set on the node by the scheduler plugin in case they don't want to override it. Now they have to write:
```python
if self.node.exit_status is not None:
from aiida.engine import ExitCode
return ExitCode(self.node.exit_status, self.node.exit_message)
```
which will then be simplified to:
```python
if self.node.exit_code:
return self.node.exit_code
```
|
1.0
|
Add the `ProcessNode.exit_code` property - This property should automatically reconstitute an `ExitCode` instance from the `exit_status` and `exit_message` attributes, if defined. This is very useful in `Parser` implementations that need to return the `ExitCode` set on the node by the scheduler plugin in case they don't want to override it. Now they have to write:
```python
if self.node.exit_status is not None:
from aiida.engine import ExitCode
return ExitCode(self.node.exit_status, self.node.exit_message)
```
which will then be simplified to:
```python
if self.node.exit_code:
return self.node.exit_code
```
|
process
|
add the processnode exit code property this property should automatically reconstitute an exitcode instance from the exit status and exit message attributes if defined this is very useful in parser implementations that need to return the exitcode set on the node by the scheduler plugin in case they don t want to override it now they have to write python if self node exit status is not none from aiida engine import exitcode return exitcode self node exit status self node exit message which will then be simplified to python if self node exit code return self node exit code
| 1
|
1,737
| 4,424,992,424
|
IssuesEvent
|
2016-08-16 14:16:50
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
opened
|
Cors call activiti
|
comp: activiti-processList comp: activiti-taskList comp: activiti/form demo app
|
The activiti integration in the demo shell doesn't work unless you open it with the following command:
open -n -a /Applications/Google\ Chrome.app/ --args --user-data-dir=/tmp/chrome_dev_session --disable-web-security --allow-running-insecure-content --new-window
|
1.0
|
Cors call activiti - The activiti integration in the demo shell doesn't work unless you open it with the following command:
open -n -a /Applications/Google\ Chrome.app/ --args --user-data-dir=/tmp/chrome_dev_session --disable-web-security --allow-running-insecure-content --new-window
|
process
|
cors call activiti the activiti integration in the demo shell doesn t work unless you open it with the following command open n a applications google chrome app args user data dir tmp chrome dev session disable web security allow running insecure content new window
| 1
|
7,537
| 10,617,386,994
|
IssuesEvent
|
2019-10-12 18:44:07
|
kubeflow/website
|
https://api.github.com/repos/kubeflow/website
|
closed
|
Implement the new docs team planning process
|
area/process kind/feature
|
Set up infrastructure (spreadsheet, mirroriing of GitHub issues). Create doc issues with the required fields for estimates etc. Build monthly sprint backlogs.
|
1.0
|
Implement the new docs team planning process - Set up infrastructure (spreadsheet, mirroriing of GitHub issues). Create doc issues with the required fields for estimates etc. Build monthly sprint backlogs.
|
process
|
implement the new docs team planning process set up infrastructure spreadsheet mirroriing of github issues create doc issues with the required fields for estimates etc build monthly sprint backlogs
| 1
|
19,836
| 26,234,799,857
|
IssuesEvent
|
2023-01-05 05:49:37
|
vesoft-inc/nebula
|
https://api.github.com/repos/vesoft-inc/nebula
|
closed
|
Graphd does not handle three-value logic properly
|
type/bug severity/major auto-sync find/automation affects/master process/done
|
**Please check the FAQ documentation before raising an issue**
<!-- Please check the [FAQ](https://docs.nebula-graph.com.cn/master/20.appendix/0.FAQ/) documentation and old issues before raising an issue in case someone has asked the same question that you are asking. -->
**Describe the bug (__required__)**
Look at the queries in Nebula below:
```txt
(root@nebula) [nebulal_gdlancer]> MATCH (v0)-[e0:Rel_1|Rel_0|Rel_3]->()-[e1]->() WHERE (id(v0) in [19, 6, 17, 18, 16, 15, 11]) and e1.Rel_1_1_Bool == false and e0.Rel_3_5_Bool is null RETURN e0.Rel_3_5_Bool, e1.Rel_1_1_Bool, e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool
+-----------------+-----------------+---------------------------------------+
| e0.Rel_3_5_Bool | e1.Rel_1_1_Bool | (e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool) |
+-----------------+-----------------+---------------------------------------+
| UNKNOWN_PROP | false | UNKNOWN_PROP |
| UNKNOWN_PROP | false | UNKNOWN_PROP |
| UNKNOWN_PROP | false | UNKNOWN_PROP |
+-----------------+-----------------+---------------------------------------+
Got 3 rows (time spent 1.675ms/18.282ms)
Fri, 18 Nov 2022 10:22:05 CST
```
We can see that `null AND false` is evaluated to be `UNKNOWN_PROP`(another form of null), while in Neo4j, `null AND false` is evaluate to `false`:
```txt
$ MATCH (v0)-[e0:Rel_1|Rel_0|Rel_3]->()-[e1]->() WHERE (v0.id in [19, 6, 17, 18, 16, 15, 11]) and e1.Rel_1_1_Bool = false and e0.Rel_3_5_Bool is null RETURN e0.Rel_3_5_Bool, e1.Rel_1_1_Bool, e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool
βββββββββββββββββββ€ββββββββββββββββββ€ββββββββββββββββββββββββββββββββββββββ
β"e0.Rel_3_5_Bool"β"e1.Rel_1_1_Bool"β"e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool"β
βββββββββββββββββββͺββββββββββββββββββͺββββββββββββββββββββββββββββββββββββββ‘
βnull βfalse βfalse β
βββββββββββββββββββΌββββββββββββββββββΌββββββββββββββββββββββββββββββββββββββ€
βnull βfalse βfalse β
βββββββββββββββββββΌββββββββββββββββββΌββββββββββββββββββββββββββββββββββββββ€
βnull βfalse βfalse β
βββββββββββββββββββ΄ββββββββββββββββββ΄ββββββββββββββββββββββββββββββββββββββ
```
That is what we expected, beacause that comply with the rules of three values logic, see [Three-valued logic](https://en.wikipedia.org/wiki/Three-valued_logic)
<!-- A clear and concise description of what the bug is. -->
**Your Environments (__required__)**
* OS: `uname -a`
* Compiler: `g++ --version` or `clang++ --version`
* CPU: `lscpu`
* Commit id (e.g. `a3ffc7d8`) 31213eafa
**How To Reproduce(__required__)**
Steps to reproduce the behavior:
1. Step 1
2. Step 2
3. Step 3
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
**Additional context**
<!-- Provide logs and configs, or any other context to trace the problem. -->
|
1.0
|
Graphd does not handle three-value logic properly - **Please check the FAQ documentation before raising an issue**
<!-- Please check the [FAQ](https://docs.nebula-graph.com.cn/master/20.appendix/0.FAQ/) documentation and old issues before raising an issue in case someone has asked the same question that you are asking. -->
**Describe the bug (__required__)**
Look at the queries in Nebula below:
```txt
(root@nebula) [nebulal_gdlancer]> MATCH (v0)-[e0:Rel_1|Rel_0|Rel_3]->()-[e1]->() WHERE (id(v0) in [19, 6, 17, 18, 16, 15, 11]) and e1.Rel_1_1_Bool == false and e0.Rel_3_5_Bool is null RETURN e0.Rel_3_5_Bool, e1.Rel_1_1_Bool, e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool
+-----------------+-----------------+---------------------------------------+
| e0.Rel_3_5_Bool | e1.Rel_1_1_Bool | (e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool) |
+-----------------+-----------------+---------------------------------------+
| UNKNOWN_PROP | false | UNKNOWN_PROP |
| UNKNOWN_PROP | false | UNKNOWN_PROP |
| UNKNOWN_PROP | false | UNKNOWN_PROP |
+-----------------+-----------------+---------------------------------------+
Got 3 rows (time spent 1.675ms/18.282ms)
Fri, 18 Nov 2022 10:22:05 CST
```
We can see that `null AND false` is evaluated to be `UNKNOWN_PROP`(another form of null), while in Neo4j, `null AND false` is evaluate to `false`:
```txt
$ MATCH (v0)-[e0:Rel_1|Rel_0|Rel_3]->()-[e1]->() WHERE (v0.id in [19, 6, 17, 18, 16, 15, 11]) and e1.Rel_1_1_Bool = false and e0.Rel_3_5_Bool is null RETURN e0.Rel_3_5_Bool, e1.Rel_1_1_Bool, e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool
βββββββββββββββββββ€ββββββββββββββββββ€ββββββββββββββββββββββββββββββββββββββ
β"e0.Rel_3_5_Bool"β"e1.Rel_1_1_Bool"β"e0.Rel_3_5_Bool AND e1.Rel_1_1_Bool"β
βββββββββββββββββββͺββββββββββββββββββͺββββββββββββββββββββββββββββββββββββββ‘
βnull βfalse βfalse β
βββββββββββββββββββΌββββββββββββββββββΌββββββββββββββββββββββββββββββββββββββ€
βnull βfalse βfalse β
βββββββββββββββββββΌββββββββββββββββββΌββββββββββββββββββββββββββββββββββββββ€
βnull βfalse βfalse β
βββββββββββββββββββ΄ββββββββββββββββββ΄ββββββββββββββββββββββββββββββββββββββ
```
That is what we expected, beacause that comply with the rules of three values logic, see [Three-valued logic](https://en.wikipedia.org/wiki/Three-valued_logic)
<!-- A clear and concise description of what the bug is. -->
**Your Environments (__required__)**
* OS: `uname -a`
* Compiler: `g++ --version` or `clang++ --version`
* CPU: `lscpu`
* Commit id (e.g. `a3ffc7d8`) 31213eafa
**How To Reproduce(__required__)**
Steps to reproduce the behavior:
1. Step 1
2. Step 2
3. Step 3
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
**Additional context**
<!-- Provide logs and configs, or any other context to trace the problem. -->
|
process
|
graphd does not handle three value logic properly please check the faq documentation before raising an issue describe the bug required look at the queries in nebula below txt root nebula match where id in and rel bool false and rel bool is null return rel bool rel bool rel bool and rel bool rel bool rel bool rel bool and rel bool unknown prop false unknown prop unknown prop false unknown prop unknown prop false unknown prop got rows time spent fri nov cst we can see that null and false is evaluated to be unknown prop another form of null while in null and false is evaluate to false txt match where id in and rel bool false and rel bool is null return rel bool rel bool rel bool and rel bool βββββββββββββββββββ€ββββββββββββββββββ€ββββββββββββββββββββββββββββββββββββββ β rel bool β rel bool β rel bool and rel bool β βββββββββββββββββββͺββββββββββββββββββͺββββββββββββββββββββββββββββββββββββββ‘ βnull βfalse βfalse β βββββββββββββββββββΌββββββββββββββββββΌββββββββββββββββββββββββββββββββββββββ€ βnull βfalse βfalse β βββββββββββββββββββΌββββββββββββββββββΌββββββββββββββββββββββββββββββββββββββ€ βnull βfalse βfalse β βββββββββββββββββββ΄ββββββββββββββββββ΄ββββββββββββββββββββββββββββββββββββββ that is what we expected beacause that comply with the rules of three values logic see your environments required os uname a compiler g version or clang version cpu lscpu commit id e g how to reproduce required steps to reproduce the behavior step step step expected behavior additional context
| 1
|
15,824
| 20,018,575,631
|
IssuesEvent
|
2022-02-01 14:27:39
|
ankidroid/Anki-Android
|
https://api.github.com/repos/ankidroid/Anki-Android
|
closed
|
Disable "html_javascript_debugging" for unit/android tests
|
Performance Good First Issue! Test process
|
This code will make tests slower, as it unnecessarily introduces IO to the card rendering pipeline
https://github.com/ankidroid/Anki-Android/blob/e55051d054a10fcf8781095d874a73ec811a08ee/AnkiDroid/src/main/java/com/ichi2/anki/AnkiDroidApp.java#L333-L335
|
1.0
|
Disable "html_javascript_debugging" for unit/android tests - This code will make tests slower, as it unnecessarily introduces IO to the card rendering pipeline
https://github.com/ankidroid/Anki-Android/blob/e55051d054a10fcf8781095d874a73ec811a08ee/AnkiDroid/src/main/java/com/ichi2/anki/AnkiDroidApp.java#L333-L335
|
process
|
disable html javascript debugging for unit android tests this code will make tests slower as it unnecessarily introduces io to the card rendering pipeline
| 1
|
7,526
| 10,599,527,917
|
IssuesEvent
|
2019-10-10 08:09:49
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
closed
|
search >filter by entities
|
2.0.8 Process bug
|
open entities with the same name in all tabs
go to search and search the name
press on filter
result : the filter give you always documents

|
1.0
|
search >filter by entities - open entities with the same name in all tabs
go to search and search the name
press on filter
result : the filter give you always documents

|
process
|
search filter by entities open entities with the same name in all tabs go to search and search the name press on filter result the filter give you always documents
| 1
|
12,146
| 14,741,333,068
|
IssuesEvent
|
2021-01-07 10:27:39
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
PK is missing when migrating from @id and @@unique to @@id
|
bug/0-needs-info kind/bug process/candidate team/migrations tech/engines topic: migrate
|
<!--
Thanks for helping us improve Prisma! π Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client.
Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports
-->
## Bug description
When migrating from a table with an `@id` field attribute (PK) and a composite `@@unique` attribute to an `@@id` attribute (instead of the `@id` and `@@unique`), the table is left without an actual PK.
## How to reproduce
Steps to reproduce the behavior:
1. Create schema.prisma file with the following model
```
model model1 {
id String @default(cuid()) @id
a String
b String
c String
@@unique([a, b, c])
}
```
2. Run prisma migrate save
3. Run prisma migrate up
4. Change the the schema.prisma file as follows:
```diff
model model1 {
- id String @default(cuid()) @id
a String
b String
c String
- @@unique([a, b, c])
+ @@id([a, b, c])
}
```
5. Run prisma migrate save
6. Run prisma migrate up
7. Open the management tool of the DB (e.g. pgAdmin) and inspect the table.
8. The table has no PK at all.
## Expected behavior
It is expected to see a composite PK on the table.
## Prisma information
This it the steps.json file of the last migration:
```json
{
"version": "0.3.14-fixed",
"steps": [
{
"tag": "DeleteField",
"model": "model1",
"field": "id"
},
{
"tag": "CreateDirective",
"location": {
"path": {
"tag": "Model",
"model": "model1"
},
"directive": "id"
}
},
{
"tag": "CreateArgument",
"location": {
"tag": "Directive",
"path": {
"tag": "Model",
"model": "model1"
},
"directive": "id"
},
"argument": "",
"value": "[a, b, c]"
},
{
"tag": "DeleteDirective",
"location": {
"path": {
"tag": "Model",
"model": "model1",
"arguments": [
{
"name": "",
"value": "[a, b, c]"
}
]
},
"directive": "unique"
}
}
]
}
```
## Environment & setup
- OS: Windows
- Database: PostgreSQL
- Node.js version: v12.16.1
- Prisma version:
<!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]-->
```
Prisma CLI version: prisma/1.34.10 (windows-x64) node-v12.16.1
```
|
1.0
|
PK is missing when migrating from @id and @@unique to @@id - <!--
Thanks for helping us improve Prisma! π Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client.
Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports
-->
## Bug description
When migrating from a table with an `@id` field attribute (PK) and a composite `@@unique` attribute to an `@@id` attribute (instead of the `@id` and `@@unique`), the table is left without an actual PK.
## How to reproduce
Steps to reproduce the behavior:
1. Create schema.prisma file with the following model
```
model model1 {
id String @default(cuid()) @id
a String
b String
c String
@@unique([a, b, c])
}
```
2. Run prisma migrate save
3. Run prisma migrate up
4. Change the the schema.prisma file as follows:
```diff
model model1 {
- id String @default(cuid()) @id
a String
b String
c String
- @@unique([a, b, c])
+ @@id([a, b, c])
}
```
5. Run prisma migrate save
6. Run prisma migrate up
7. Open the management tool of the DB (e.g. pgAdmin) and inspect the table.
8. The table has no PK at all.
## Expected behavior
It is expected to see a composite PK on the table.
## Prisma information
This it the steps.json file of the last migration:
```json
{
"version": "0.3.14-fixed",
"steps": [
{
"tag": "DeleteField",
"model": "model1",
"field": "id"
},
{
"tag": "CreateDirective",
"location": {
"path": {
"tag": "Model",
"model": "model1"
},
"directive": "id"
}
},
{
"tag": "CreateArgument",
"location": {
"tag": "Directive",
"path": {
"tag": "Model",
"model": "model1"
},
"directive": "id"
},
"argument": "",
"value": "[a, b, c]"
},
{
"tag": "DeleteDirective",
"location": {
"path": {
"tag": "Model",
"model": "model1",
"arguments": [
{
"name": "",
"value": "[a, b, c]"
}
]
},
"directive": "unique"
}
}
]
}
```
## Environment & setup
- OS: Windows
- Database: PostgreSQL
- Node.js version: v12.16.1
- Prisma version:
<!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]-->
```
Prisma CLI version: prisma/1.34.10 (windows-x64) node-v12.16.1
```
|
process
|
pk is missing when migrating from id and unique to id thanks for helping us improve prisma π please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description when migrating from a table with an id field attribute pk and a composite unique attribute to an id attribute instead of the id and unique the table is left without an actual pk how to reproduce steps to reproduce the behavior create schema prisma file with the following model model id string default cuid id a string b string c string unique run prisma migrate save run prisma migrate up change the the schema prisma file as follows diff model id string default cuid id a string b string c string unique id run prisma migrate save run prisma migrate up open the management tool of the db e g pgadmin and inspect the table the table has no pk at all expected behavior it is expected to see a composite pk on the table prisma information this it the steps json file of the last migration json version fixed steps tag deletefield model field id tag createdirective location path tag model model directive id tag createargument location tag directive path tag model model directive id argument value tag deletedirective location path tag model model arguments name value directive unique environment setup os windows database postgresql node js version prisma version prisma cli version prisma windows node
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.