Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
171,439
27,119,847,218
IssuesEvent
2023-02-15 21:48:18
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
API deprecate date_parser, add date_format
Timeseries API Design IO CSV
**TLDR** the conversation here goes on for a bit, but to summarise, the suggestion is: - deprecate `date_parser`, because it always hurts performance (counter examples welcome!) - add `date_format`, because that can boost performance - for anything else, amend docs to make clear that users should first in the data as `object`, and then apply their parsing Performance-wise, this would only be an improvement to the status quo ------ As far as I can tell, `date_parser` is a net negative and only ever slows things down In the best case, it only results in a slight degradation: ```python timestamp_format = '%Y-%d-%m %H:%M:%S' date_index = pd.date_range(start='1900', end='2000') dates_df = date_index.strftime(timestamp_format).to_frame(name='ts_col') data = dates_df.to_csv() ``` ```python In [6]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: date_parser=lambda x: pd.to_datetime(x, format=timestamp_format), ...: parse_dates=['ts_col'] ...: ) ...: ...: 111 ms ± 3.02 ms per loop (mean ± std. dev. of 7 runs, 10 loops each) In [7]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: #date_parser=lambda x: pd.to_datetime(x, format=timestamp_format), ...: #parse_dates=['ts_col'] ...: ) ...: df['ts_col'] = pd.to_datetime(df['ts_col'], format=timestamp_format) ...: ...: 75.8 ms ± 1.98 ms per loop (mean ± std. dev. of 7 runs, 10 loops each) ``` Parsing element-by-element is also slower than just using `.apply`: ```python In [21]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: #date_parser=lambda x: pd.to_datetime(x, format=timestamp_format), ...: #parse_dates=['ts_col'] ...: ) ...: df['ts_col'].apply(lambda x: du_parse(x, dayfirst=True)) ...: ...: 1.13 s ± 33.4 ms per loop (mean ± std. dev. of 7 runs, 1 loop each) In [22]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: date_parser=lambda x: du_parse(x, dayfirst=True), ...: parse_dates=['ts_col'] ...: ) ...: ...: 1.19 s ± 3.43 ms per loop (mean ± std. dev. of 7 runs, 1 loop each) ``` In the worst case, it results in **65x performance degradation**, see https://github.com/pandas-dev/pandas/pull/50586#issuecomment-1373473579 (and this gets way worse for larger datasets) My suggestion is: - deprecate `date_parser` - introduce `date_format`, which _would_ actually deliver a performance improvement: ```python In [1]: timestamp_format = '%Y-%d-%m %H:%M:%S' ...: ...: date_index = pd.date_range(start='1900', end='2000') ...: ...: dates_df = date_index.strftime(timestamp_format).to_frame(name='ts_col') ...: data = dates_df.to_csv() In [2]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: date_format=timestamp_format, ...: parse_dates=['ts_col'] ...: ) ...: ...: 19.5 ms ± 397 µs per loop (mean ± std. dev. of 7 runs, 10 loops each) ``` That's over 3 times as fast!
1.0
API deprecate date_parser, add date_format - **TLDR** the conversation here goes on for a bit, but to summarise, the suggestion is: - deprecate `date_parser`, because it always hurts performance (counter examples welcome!) - add `date_format`, because that can boost performance - for anything else, amend docs to make clear that users should first in the data as `object`, and then apply their parsing Performance-wise, this would only be an improvement to the status quo ------ As far as I can tell, `date_parser` is a net negative and only ever slows things down In the best case, it only results in a slight degradation: ```python timestamp_format = '%Y-%d-%m %H:%M:%S' date_index = pd.date_range(start='1900', end='2000') dates_df = date_index.strftime(timestamp_format).to_frame(name='ts_col') data = dates_df.to_csv() ``` ```python In [6]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: date_parser=lambda x: pd.to_datetime(x, format=timestamp_format), ...: parse_dates=['ts_col'] ...: ) ...: ...: 111 ms ± 3.02 ms per loop (mean ± std. dev. of 7 runs, 10 loops each) In [7]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: #date_parser=lambda x: pd.to_datetime(x, format=timestamp_format), ...: #parse_dates=['ts_col'] ...: ) ...: df['ts_col'] = pd.to_datetime(df['ts_col'], format=timestamp_format) ...: ...: 75.8 ms ± 1.98 ms per loop (mean ± std. dev. of 7 runs, 10 loops each) ``` Parsing element-by-element is also slower than just using `.apply`: ```python In [21]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: #date_parser=lambda x: pd.to_datetime(x, format=timestamp_format), ...: #parse_dates=['ts_col'] ...: ) ...: df['ts_col'].apply(lambda x: du_parse(x, dayfirst=True)) ...: ...: 1.13 s ± 33.4 ms per loop (mean ± std. dev. of 7 runs, 1 loop each) In [22]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: date_parser=lambda x: du_parse(x, dayfirst=True), ...: parse_dates=['ts_col'] ...: ) ...: ...: 1.19 s ± 3.43 ms per loop (mean ± std. dev. of 7 runs, 1 loop each) ``` In the worst case, it results in **65x performance degradation**, see https://github.com/pandas-dev/pandas/pull/50586#issuecomment-1373473579 (and this gets way worse for larger datasets) My suggestion is: - deprecate `date_parser` - introduce `date_format`, which _would_ actually deliver a performance improvement: ```python In [1]: timestamp_format = '%Y-%d-%m %H:%M:%S' ...: ...: date_index = pd.date_range(start='1900', end='2000') ...: ...: dates_df = date_index.strftime(timestamp_format).to_frame(name='ts_col') ...: data = dates_df.to_csv() In [2]: %%timeit ...: df = pd.read_csv(io.StringIO(data), ...: date_format=timestamp_format, ...: parse_dates=['ts_col'] ...: ) ...: ...: 19.5 ms ± 397 µs per loop (mean ± std. dev. of 7 runs, 10 loops each) ``` That's over 3 times as fast!
non_process
api deprecate date parser add date format tldr the conversation here goes on for a bit but to summarise the suggestion is deprecate date parser because it always hurts performance counter examples welcome add date format because that can boost performance for anything else amend docs to make clear that users should first in the data as object and then apply their parsing performance wise this would only be an improvement to the status quo as far as i can tell date parser is a net negative and only ever slows things down in the best case it only results in a slight degradation python timestamp format y d m h m s date index pd date range start end dates df date index strftime timestamp format to frame name ts col data dates df to csv python in timeit df pd read csv io stringio data date parser lambda x pd to datetime x format timestamp format parse dates ms ± ms per loop mean ± std dev of runs loops each in timeit df pd read csv io stringio data date parser lambda x pd to datetime x format timestamp format parse dates df pd to datetime df format timestamp format ms ± ms per loop mean ± std dev of runs loops each parsing element by element is also slower than just using apply python in timeit df pd read csv io stringio data date parser lambda x pd to datetime x format timestamp format parse dates df apply lambda x du parse x dayfirst true s ± ms per loop mean ± std dev of runs loop each in timeit df pd read csv io stringio data date parser lambda x du parse x dayfirst true parse dates s ± ms per loop mean ± std dev of runs loop each in the worst case it results in performance degradation see and this gets way worse for larger datasets my suggestion is deprecate date parser introduce date format which would actually deliver a performance improvement python in timestamp format y d m h m s date index pd date range start end dates df date index strftime timestamp format to frame name ts col data dates df to csv in timeit df pd read csv io stringio data date format timestamp format parse dates ms ± µs per loop mean ± std dev of runs loops each that s over times as fast
0
270,882
23,545,252,395
IssuesEvent
2022-08-21 02:16:48
puppetlabs/pdk
https://api.github.com/repos/puppetlabs/pdk
closed
Long time from starting pdk test unit until unit tests start
needs information pdk-test no-issue-activity
**Describe the bug** When running `pdk test unit` it takes a very long time where PDK is doing something prior the unit tests start **To Reproduce** run `time pdk test unit` **Expected behavior** Unit tests start immediately after running `pdk test unit` **Additional context** Installed via DEB Package v1.13.0 on Debian 9
1.0
Long time from starting pdk test unit until unit tests start - **Describe the bug** When running `pdk test unit` it takes a very long time where PDK is doing something prior the unit tests start **To Reproduce** run `time pdk test unit` **Expected behavior** Unit tests start immediately after running `pdk test unit` **Additional context** Installed via DEB Package v1.13.0 on Debian 9
non_process
long time from starting pdk test unit until unit tests start describe the bug when running pdk test unit it takes a very long time where pdk is doing something prior the unit tests start to reproduce run time pdk test unit expected behavior unit tests start immediately after running pdk test unit additional context installed via deb package on debian
0
166,965
6,329,306,251
IssuesEvent
2017-07-26 02:19:58
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Timers are crashing (or something else is that is causing temp_visuals to not delete) consistently nearing 2 hours into the round
Bug Priority: High
For the past few days every time it's nearing 14:00 shift time shit just starts going wack. What happened to the robust MC that can handle everything?
1.0
Timers are crashing (or something else is that is causing temp_visuals to not delete) consistently nearing 2 hours into the round - For the past few days every time it's nearing 14:00 shift time shit just starts going wack. What happened to the robust MC that can handle everything?
non_process
timers are crashing or something else is that is causing temp visuals to not delete consistently nearing hours into the round for the past few days every time it s nearing shift time shit just starts going wack what happened to the robust mc that can handle everything
0
5,988
2,996,822,144
IssuesEvent
2015-07-23 00:47:03
swagger-api/swagger-js
https://api.github.com/repos/swagger-api/swagger-js
closed
adding clientAuthorization broken
Documentation
The example in readme.md for adding a new authorization token is busted. Using the boilerplate code plus `client.clientAuthorizations.add("apiKey", new client.ApiKeyAuthorization("api_key","special-key","query"));` results in a TypeError: `client.clientAuthorizations.add("api_key", ...` ` ^` (carat under add) `TypeError: Cannot call method 'add' of undefined` I got the authorization working using the constructor method listed in Feature request #146, but the readme page still references the old method (that appears to not work anymore).
1.0
adding clientAuthorization broken - The example in readme.md for adding a new authorization token is busted. Using the boilerplate code plus `client.clientAuthorizations.add("apiKey", new client.ApiKeyAuthorization("api_key","special-key","query"));` results in a TypeError: `client.clientAuthorizations.add("api_key", ...` ` ^` (carat under add) `TypeError: Cannot call method 'add' of undefined` I got the authorization working using the constructor method listed in Feature request #146, but the readme page still references the old method (that appears to not work anymore).
non_process
adding clientauthorization broken the example in readme md for adding a new authorization token is busted using the boilerplate code plus client clientauthorizations add apikey new client apikeyauthorization api key special key query results in a typeerror client clientauthorizations add api key carat under add typeerror cannot call method add of undefined i got the authorization working using the constructor method listed in feature request but the readme page still references the old method that appears to not work anymore
0
8,772
23,421,737,082
IssuesEvent
2022-08-13 20:06:55
nix-community/dream2nix
https://api.github.com/repos/nix-community/dream2nix
closed
per-subsystem subsystemInfo definition/parser?
enhancement architecture
It feels weird to let subsystemInfo be handled by translator only, especially since there can be many translators, passing settings to the builder is arbitrary, and I guess the discoverer can't see them? How about a per-subsystem settings module that defines the flags and optionally transforms them? Then they could also be made available to the overrides?
1.0
per-subsystem subsystemInfo definition/parser? - It feels weird to let subsystemInfo be handled by translator only, especially since there can be many translators, passing settings to the builder is arbitrary, and I guess the discoverer can't see them? How about a per-subsystem settings module that defines the flags and optionally transforms them? Then they could also be made available to the overrides?
non_process
per subsystem subsysteminfo definition parser it feels weird to let subsysteminfo be handled by translator only especially since there can be many translators passing settings to the builder is arbitrary and i guess the discoverer can t see them how about a per subsystem settings module that defines the flags and optionally transforms them then they could also be made available to the overrides
0
12,825
15,210,184,604
IssuesEvent
2021-02-17 06:58:38
gfx-rs/naga
https://api.github.com/repos/gfx-rs/naga
opened
Global usage flags are not correctly derived
area: processing kind: bug
Related to #486 I don't think we correctly identify things like `buffer.array[a+4] = b`. Here, the LHS expression is `Access`, not the global itself. We need to make it so `buffer` is identified as being written to, and `a` with `b` are read.
1.0
Global usage flags are not correctly derived - Related to #486 I don't think we correctly identify things like `buffer.array[a+4] = b`. Here, the LHS expression is `Access`, not the global itself. We need to make it so `buffer` is identified as being written to, and `a` with `b` are read.
process
global usage flags are not correctly derived related to i don t think we correctly identify things like buffer array b here the lhs expression is access not the global itself we need to make it so buffer is identified as being written to and a with b are read
1
12,542
14,974,454,200
IssuesEvent
2021-01-28 03:35:48
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] Resources > PDFs uploaded in resources are broken on accessing for the first time
Bug P1 Process: Dev Process: Fixed iOS
**Describe the bug** PDFs uploaded in resources are broken on accessing it for the first time **To Reproduce** Steps to reproduce the behavior: 1. Join the study successfully 2. Go to 'Resources' 3. Click on any resources having PDF 4. Observe the format 5. Navigate back and again click the same PDF 6. Observe PDF is loaded without any issue **Actual behavior** PDFs are broken when accessed for first time **Expected behavior** PDFs should load properly Refer attached video https://user-images.githubusercontent.com/60386291/103407200-bbe3ce80-4b83-11eb-8c1f-e12744267b9a.MP4
2.0
[iOS] Resources > PDFs uploaded in resources are broken on accessing for the first time - **Describe the bug** PDFs uploaded in resources are broken on accessing it for the first time **To Reproduce** Steps to reproduce the behavior: 1. Join the study successfully 2. Go to 'Resources' 3. Click on any resources having PDF 4. Observe the format 5. Navigate back and again click the same PDF 6. Observe PDF is loaded without any issue **Actual behavior** PDFs are broken when accessed for first time **Expected behavior** PDFs should load properly Refer attached video https://user-images.githubusercontent.com/60386291/103407200-bbe3ce80-4b83-11eb-8c1f-e12744267b9a.MP4
process
resources pdfs uploaded in resources are broken on accessing for the first time describe the bug pdfs uploaded in resources are broken on accessing it for the first time to reproduce steps to reproduce the behavior join the study successfully go to resources click on any resources having pdf observe the format navigate back and again click the same pdf observe pdf is loaded without any issue actual behavior pdfs are broken when accessed for first time expected behavior pdfs should load properly refer attached video
1
59,700
8,376,996,575
IssuesEvent
2018-10-05 22:02:25
opal/opal
https://api.github.com/repos/opal/opal
closed
Browser support
discuss documentation
> On an unrelated subject (well, maybe not so unrelated), I believe the claim that Opal runs on IE 6 is no longer true. I have seen things like Object.keys used in corelib (I think in the Array or Hash implementation), and that only became available in IE 9 I think. Not sure what if anything needs to be done about that, but wanted to hear your thoughts on what browsers should (ideally vs practically) be supported and how to make sure they are as work on Opal continues. > — @vais in #726 cc @adambeynon @meh
1.0
Browser support - > On an unrelated subject (well, maybe not so unrelated), I believe the claim that Opal runs on IE 6 is no longer true. I have seen things like Object.keys used in corelib (I think in the Array or Hash implementation), and that only became available in IE 9 I think. Not sure what if anything needs to be done about that, but wanted to hear your thoughts on what browsers should (ideally vs practically) be supported and how to make sure they are as work on Opal continues. > — @vais in #726 cc @adambeynon @meh
non_process
browser support on an unrelated subject well maybe not so unrelated i believe the claim that opal runs on ie is no longer true i have seen things like object keys used in corelib i think in the array or hash implementation and that only became available in ie i think not sure what if anything needs to be done about that but wanted to hear your thoughts on what browsers should ideally vs practically be supported and how to make sure they are as work on opal continues — vais in cc adambeynon meh
0
1,087
3,550,200,172
IssuesEvent
2016-01-20 21:02:22
nodejs/node
https://api.github.com/repos/nodejs/node
closed
hrtime regression
process
I have the feeling https://github.com/nodejs/node/commit/89f056bdf323a77a48997f14196b41ddf76f90b1 introduced a regression. The nanoseconds might return negative values if you compare the actual value with the one before. This does not occur in 5.3 or before. ``` > process.hrtime(t) [ 45, 373756947 ] > process.hrtime(t) [ 46, -1831860 ] ``` Easy to reproduce with: ```js var t = process.hrtime(); while(t[1] > 0) t = process.hrtime(t); ```
1.0
hrtime regression - I have the feeling https://github.com/nodejs/node/commit/89f056bdf323a77a48997f14196b41ddf76f90b1 introduced a regression. The nanoseconds might return negative values if you compare the actual value with the one before. This does not occur in 5.3 or before. ``` > process.hrtime(t) [ 45, 373756947 ] > process.hrtime(t) [ 46, -1831860 ] ``` Easy to reproduce with: ```js var t = process.hrtime(); while(t[1] > 0) t = process.hrtime(t); ```
process
hrtime regression i have the feeling introduced a regression the nanoseconds might return negative values if you compare the actual value with the one before this does not occur in or before process hrtime t process hrtime t easy to reproduce with js var t process hrtime while t t process hrtime t
1
6,582
9,661,750,759
IssuesEvent
2019-05-20 18:53:45
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
opened
Transcript: Default radio button if only one
Apply Process State Dept.
Who: Internship applicants What: Default transcript selection if only one uploaded Why: To avoid confusion Acceptance Criteria: - When an internship applicant selects to upload a transcript and then selects "Refresh", default the radio button to select the transcript IF there was only one uploaded.
1.0
Transcript: Default radio button if only one - Who: Internship applicants What: Default transcript selection if only one uploaded Why: To avoid confusion Acceptance Criteria: - When an internship applicant selects to upload a transcript and then selects "Refresh", default the radio button to select the transcript IF there was only one uploaded.
process
transcript default radio button if only one who internship applicants what default transcript selection if only one uploaded why to avoid confusion acceptance criteria when an internship applicant selects to upload a transcript and then selects refresh default the radio button to select the transcript if there was only one uploaded
1
14,667
17,786,803,265
IssuesEvent
2021-08-31 12:05:55
googleapis/google-api-python-client
https://api.github.com/repos/googleapis/google-api-python-client
reopened
Dependency Dashboard
type: process
This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Other Branches These updates are pending. To force PRs open, click the checkbox below. - [ ] <!-- other-branch=renovate/actions-github-script-4.x -->chore(deps): update actions/github-script action to v4.1.0 ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/google-api-python-client-2.x -->[chore(deps): update dependency google-api-python-client to v2.19.0](../pull/1507) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Other Branches These updates are pending. To force PRs open, click the checkbox below. - [ ] <!-- other-branch=renovate/actions-github-script-4.x -->chore(deps): update actions/github-script action to v4.1.0 ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/google-api-python-client-2.x -->[chore(deps): update dependency google-api-python-client to v2.19.0](../pull/1507) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue provides visibility into renovate updates and their statuses other branches these updates are pending to force prs open click the checkbox below chore deps update actions github script action to open these updates have all been created already click a checkbox below to force a retry rebase of any pull check this box to trigger a request for renovate to run again on this repository
1
51,929
12,828,632,103
IssuesEvent
2020-07-06 20:56:53
google/blockly
https://api.github.com/repos/google/blockly
opened
Updata metadata size as part of release script
component: build process type: feature request
**Is your feature request related to a problem? Please describe.** We have to update the metadata size for every release. <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> **Describe the solution you'd like** Change npm run recompile -> npm run recompile:bump & npm run recompile:all npm run recompile:all will bump the version and also update the check_metadata.sh file with the current file size. This will be run only for major releases and not for patches. <!-- A clear and concise description of what you want to happen. --> **Describe alternatives you've considered** Doing it manually : ( <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
1.0
Updata metadata size as part of release script - **Is your feature request related to a problem? Please describe.** We have to update the metadata size for every release. <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> **Describe the solution you'd like** Change npm run recompile -> npm run recompile:bump & npm run recompile:all npm run recompile:all will bump the version and also update the check_metadata.sh file with the current file size. This will be run only for major releases and not for patches. <!-- A clear and concise description of what you want to happen. --> **Describe alternatives you've considered** Doing it manually : ( <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
non_process
updata metadata size as part of release script is your feature request related to a problem please describe we have to update the metadata size for every release describe the solution you d like change npm run recompile npm run recompile bump npm run recompile all npm run recompile all will bump the version and also update the check metadata sh file with the current file size this will be run only for major releases and not for patches describe alternatives you ve considered doing it manually additional context
0
280,702
24,324,753,152
IssuesEvent
2022-09-30 13:53:54
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: unoptimized-query-oracle/disable-rules=half failed
C-test-failure O-robot O-roachtest T-sql-queries branch-release-22.2
roachtest.unoptimized-query-oracle/disable-rules=half [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6704805?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6704805?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=half) on release-22.2 @ [c69571de022f48d95d76fb39e378cc9ab9a30afe](https://github.com/cockroachdb/cockroach/commits/c69571de022f48d95d76fb39e378cc9ab9a30afe): ``` |    ",NULL", |    }, ""), |    strings.Join({ |    ",1664543667946134844.0000000000,,-5", | -  ".", |    "954290679578412342", | -  "E+35", | +  "00000000000000000", |    ",{},105,àá,0,0,[[[]], {\"a\": {\",}3^``\": {}, \"Hu-;@EAJWDa\": true}", |    ``, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": {}}, [], "a"],``, |    "NULL,NULL", |    }, ""), |    strings.Join({ |    ",1664543669528713175.0000000000,,-5", | -  ".", |    "954290679578412342", | -  "E+35", | +  "00000000000000000", |    ",{\"test\": \"json\"},105,àá,0,0,[[[]], {\"a\": {\",}3^``\": {}, \"Hu-;@", |    ``EAJWDa": true}, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": ``, |    ``{}}, [], "a"],NULL,0``, |    }, ""), |    strings.Join({ |    ",1664543670755270935.0000000000,,-5", | -  ".", |    "954290679578412342", | -  "E+35", | +  "00000000000000000", |    ",NULL,105,àá,0,0,[[[]], {\"a\": {\",}3^``\": {}, \"Hu-;@EAJWDa\": tru", |    ``e}, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": {}}, [], "a"``, |    "],NULL,NULL", |    }, ""), |   } | sql: SELECT | tab_9739.col2_9 AS col_29539, | tab_9739.crdb_internal_mvcc_timestamp AS col_29540, | tab_9739.col2_9 AS col_29541, | ((-5.954290679578412342E+35):::DECIMAL::DECIMAL + 0:::DECIMAL::DECIMAL)::DECIMAL AS col_29542, | tab_9739.col2_8 AS col_29543, | tab_9739.tableoid AS col_29544, | '\xc3a0c3a1':::BYTES AS col_29545, | 0:::OID AS col_29546, | 0:::OID AS col_29547, | '[[[]], {"a": {",}3^``": {}, "Hu-;@EAJWDa": true}, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": {}}, [], "a"]':::JSONB | AS col_29548, | tab_9739.col2_2 AS col_29549, | tab_9739.col2_6 AS col_29550 | FROM | defaultdb.public.table2@[0] AS tab_9739 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=half.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: unoptimized-query-oracle/disable-rules=half failed - roachtest.unoptimized-query-oracle/disable-rules=half [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6704805?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6704805?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=half) on release-22.2 @ [c69571de022f48d95d76fb39e378cc9ab9a30afe](https://github.com/cockroachdb/cockroach/commits/c69571de022f48d95d76fb39e378cc9ab9a30afe): ``` |    ",NULL", |    }, ""), |    strings.Join({ |    ",1664543667946134844.0000000000,,-5", | -  ".", |    "954290679578412342", | -  "E+35", | +  "00000000000000000", |    ",{},105,àá,0,0,[[[]], {\"a\": {\",}3^``\": {}, \"Hu-;@EAJWDa\": true}", |    ``, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": {}}, [], "a"],``, |    "NULL,NULL", |    }, ""), |    strings.Join({ |    ",1664543669528713175.0000000000,,-5", | -  ".", |    "954290679578412342", | -  "E+35", | +  "00000000000000000", |    ",{\"test\": \"json\"},105,àá,0,0,[[[]], {\"a\": {\",}3^``\": {}, \"Hu-;@", |    ``EAJWDa": true}, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": ``, |    ``{}}, [], "a"],NULL,0``, |    }, ""), |    strings.Join({ |    ",1664543670755270935.0000000000,,-5", | -  ".", |    "954290679578412342", | -  "E+35", | +  "00000000000000000", |    ",NULL,105,àá,0,0,[[[]], {\"a\": {\",}3^``\": {}, \"Hu-;@EAJWDa\": tru", |    ``e}, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": {}}, [], "a"``, |    "],NULL,NULL", |    }, ""), |   } | sql: SELECT | tab_9739.col2_9 AS col_29539, | tab_9739.crdb_internal_mvcc_timestamp AS col_29540, | tab_9739.col2_9 AS col_29541, | ((-5.954290679578412342E+35):::DECIMAL::DECIMAL + 0:::DECIMAL::DECIMAL)::DECIMAL AS col_29542, | tab_9739.col2_8 AS col_29543, | tab_9739.tableoid AS col_29544, | '\xc3a0c3a1':::BYTES AS col_29545, | 0:::OID AS col_29546, | 0:::OID AS col_29547, | '[[[]], {"a": {",}3^``": {}, "Hu-;@EAJWDa": true}, "jduN9Sf>": 2.8114104318075186, "{h(XYi|CHoFu": {}}, [], "a"]':::JSONB | AS col_29548, | tab_9739.col2_2 AS col_29549, | tab_9739.col2_6 AS col_29550 | FROM | defaultdb.public.table2@[0] AS tab_9739 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=half.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_process
roachtest unoptimized query oracle disable rules half failed roachtest unoptimized query oracle disable rules half with on release    null       strings join           e      àá a hu eajwda true    h xyi chofu a    null null       strings join           e      test json àá a hu    eajwda true h xyi chofu    a null       strings join           e      null àá a hu eajwda tru    e h xyi chofu a    null null       sql select tab as col tab crdb internal mvcc timestamp as col tab as col decimal decimal decimal decimal decimal as col tab as col tab tableoid as col bytes as col oid as col oid as col a hu eajwda true h xyi chofu a jsonb as col tab as col tab as col from defaultdb public as tab error types withstack withstack errutil withprefix withstack withstack errutil leaferror parameters roachtest cloud gce roachtest cpu roachtest ssd help see see cc cockroachdb sql queries
0
19,397
25,539,287,997
IssuesEvent
2022-11-29 14:16:30
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Preprocessor function multi_model_statistics does not work if no time coordinate or only single time point in data
bug preprocessor
Because the code assumes a time coordinate is present and does `cube.coord('time').guess_bounds()`
1.0
Preprocessor function multi_model_statistics does not work if no time coordinate or only single time point in data - Because the code assumes a time coordinate is present and does `cube.coord('time').guess_bounds()`
process
preprocessor function multi model statistics does not work if no time coordinate or only single time point in data because the code assumes a time coordinate is present and does cube coord time guess bounds
1
4,963
7,806,051,135
IssuesEvent
2018-06-11 12:59:15
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Per connection rate limit
ADMIN CONNECTION POOL QUERY PROCESSOR
We need to define a way to throttle bandwidth usage per connection/query, client side.
1.0
Per connection rate limit - We need to define a way to throttle bandwidth usage per connection/query, client side.
process
per connection rate limit we need to define a way to throttle bandwidth usage per connection query client side
1
414
7,703,837,749
IssuesEvent
2018-05-21 09:53:35
DrewAPicture/ensemble
https://api.github.com/repos/DrewAPicture/ensemble
opened
Term query support for Unit Directors
::People ::Taxonomy Enhancement
Like contests and venues, Unit Directors (users) are connected with various taxonomies, but most notably the Units taxonomy. It would be nice if there were a way to bridge between directors and the terms assigned to them, even if it means going outside of the core API pathways to accomplish it.
1.0
Term query support for Unit Directors - Like contests and venues, Unit Directors (users) are connected with various taxonomies, but most notably the Units taxonomy. It would be nice if there were a way to bridge between directors and the terms assigned to them, even if it means going outside of the core API pathways to accomplish it.
non_process
term query support for unit directors like contests and venues unit directors users are connected with various taxonomies but most notably the units taxonomy it would be nice if there were a way to bridge between directors and the terms assigned to them even if it means going outside of the core api pathways to accomplish it
0
4,769
3,882,931,791
IssuesEvent
2016-04-13 12:02:37
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
20627241: Swift.LazySequence should have a .first property
classification:ui/usability reproducible:always status:open
#### Description Summary: LazySequence should have a `first` property of type `S.Generator.Element?`. Steps to Reproduce: 1. Try to find the first element matching a given criterion: return lazy(start…end).filter { $0 % 2 == 0 }.first Expected Results: This code, or something very similar, works fine. Actual Results: There is no first property, method, or function that can be used on a LazySequence. The closest equivalent is to use its generator directly, capturing it in a mutable variable first: var gen = lazy(start…end).filter { $0 % 2 == 0 }.generate() return gen.next() Notes: Another alternative would be to add a first() function that takes a SequenceType, but this would be uglier and also would consume elements from one-time-through sequences, which might be unexpected. Calling `first` twice on a LazySequence might run the `filter` method’s block twice. That would be fine. - Product Version: Xcode 6.3 (6D570) Created: 2015-04-21T01:02:06.653630 Originated: 2015-04-20T18:01:00 Open Radar Link: http://www.openradar.me/20627241
True
20627241: Swift.LazySequence should have a .first property - #### Description Summary: LazySequence should have a `first` property of type `S.Generator.Element?`. Steps to Reproduce: 1. Try to find the first element matching a given criterion: return lazy(start…end).filter { $0 % 2 == 0 }.first Expected Results: This code, or something very similar, works fine. Actual Results: There is no first property, method, or function that can be used on a LazySequence. The closest equivalent is to use its generator directly, capturing it in a mutable variable first: var gen = lazy(start…end).filter { $0 % 2 == 0 }.generate() return gen.next() Notes: Another alternative would be to add a first() function that takes a SequenceType, but this would be uglier and also would consume elements from one-time-through sequences, which might be unexpected. Calling `first` twice on a LazySequence might run the `filter` method’s block twice. That would be fine. - Product Version: Xcode 6.3 (6D570) Created: 2015-04-21T01:02:06.653630 Originated: 2015-04-20T18:01:00 Open Radar Link: http://www.openradar.me/20627241
non_process
swift lazysequence should have a first property description summary lazysequence should have a first property of type s generator element steps to reproduce try to find the first element matching a given criterion return lazy start…end filter first expected results this code or something very similar works fine actual results there is no first property method or function that can be used on a lazysequence the closest equivalent is to use its generator directly capturing it in a mutable variable first var gen lazy start…end filter generate return gen next notes another alternative would be to add a first function that takes a sequencetype but this would be uglier and also would consume elements from one time through sequences which might be unexpected calling first twice on a lazysequence might run the filter method’s block twice that would be fine product version xcode created originated open radar link
0
6,978
10,129,759,576
IssuesEvent
2019-08-01 15:25:34
ptddatateam/PublicTransitDataReportingPortal
https://api.github.com/repos/ptddatateam/PublicTransitDataReportingPortal
opened
Prep for meeting to discuss Summary of Public Transportation Vision
process
Prep for meeting to discuss Summary of Public Transportation Vision
1.0
Prep for meeting to discuss Summary of Public Transportation Vision - Prep for meeting to discuss Summary of Public Transportation Vision
process
prep for meeting to discuss summary of public transportation vision prep for meeting to discuss summary of public transportation vision
1
816,133
30,590,165,097
IssuesEvent
2023-07-21 16:19:34
Memmy-App/memmy
https://api.github.com/repos/Memmy-App/memmy
closed
Minor Traverse Tweaks: Favorite and Subscriptions Counts and Padding.
enhancement medium priority visual
**Describe the solution you'd like** Beside the Favorites and Subscriptions titles I would like a count of number of favorites and subscriptions. I would also like the padding between Subscriptions and the post above to match the padding between the search box and favorite title. Looks cramped. **Additional context** ![image](https://github.com/Memmy-App/memmy/assets/19845100/a11aaac1-d866-4065-8b76-055de9bdfe4f) Mock-up done on my phone, sorry.
1.0
Minor Traverse Tweaks: Favorite and Subscriptions Counts and Padding. - **Describe the solution you'd like** Beside the Favorites and Subscriptions titles I would like a count of number of favorites and subscriptions. I would also like the padding between Subscriptions and the post above to match the padding between the search box and favorite title. Looks cramped. **Additional context** ![image](https://github.com/Memmy-App/memmy/assets/19845100/a11aaac1-d866-4065-8b76-055de9bdfe4f) Mock-up done on my phone, sorry.
non_process
minor traverse tweaks favorite and subscriptions counts and padding describe the solution you d like beside the favorites and subscriptions titles i would like a count of number of favorites and subscriptions i would also like the padding between subscriptions and the post above to match the padding between the search box and favorite title looks cramped additional context mock up done on my phone sorry
0
343,230
30,653,536,342
IssuesEvent
2023-07-25 10:29:20
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix paddle_math.test_paddle_atanh
Sub Task Failing Test Paddle Frontend
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix paddle_math.test_paddle_atanh - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5522136127"><img src=https://img.shields.io/badge/-success-success></a>
non_process
fix paddle math test paddle atanh tensorflow a href src jax a href src paddle a href src torch a href src numpy a href src
0
2,628
5,400,490,947
IssuesEvent
2017-02-27 22:10:22
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Non admin users can't see Questions that are using GA as data source
Bug Database/Google Analytics Query Processor
### Bug Regular User can't access Questions and Cards using data from GoogleAnalytics database but they can ask a question and get data as long as they are not saving it. - Browser and the version: Chrome 55.0.2883.95 - Operating system: macOS Sierra - Databases: MongoDB, GoogleAnalytics - Metabase version: 0.22 - Metabase hosting environment: Docker - Metabase internal database: (default) #### Steps to reproduce - User is not admin - Create new question using Google Analytics as DB source - Get answer - Save the question - Get the permission error : <img width="672" alt="capture d ecran 2017-01-13 a 11 13 59" src="https://cloud.githubusercontent.com/assets/60847/21926465/183da5e8-d982-11e6-8ec4-7808b4c7da09.png"> This is not affecting admin users #### Logs ``` 01-13 10:21:23 WARN models.card :: Error getting permissions for card: Illegal clause (no matching fn found): :metric ["query_processor.expand$fn_for_token.invokeStatic(expand.clj:442)" "query_processor.expand$fn_for_token.invoke(expand.clj:435)" "query_processor.expand$fn__21383$expand_ql_sexpr__21388$fn__21392.invoke(expand.clj:452)" "query_processor.expand$fn__21383$expand_ql_sexpr__21388.invoke(expand.clj:444)" "query_processor.expand$walk_expand_ql_sexprs.invokeStatic(expand.clj:460)" "query_processor.expand$walk_expand_ql_sexprs.invoke(expand.clj:455)" "query_processor.expand$walk_expand_ql_sexprs.invokeStatic(expand.clj:460)" "query_processor.expand$walk_expand_ql_sexprs.invoke(expand.clj:455)" "query_processor.expand$fn__21428$expand_inner__21433$fn__21434.invoke(expand.clj:468)" "query_processor.expand$fn__21428$expand_inner__21433.invoke(expand.clj:464)" "query_processor.expand$expand.invokeStatic(expand.clj:493)" "query_processor.expand$expand.invoke(expand.clj:479)" "query_processor$expand_resolve.invokeStatic(query_processor.clj:178)" "query_processor$expand_resolve.invoke(query_processor.clj:171)" "models.card$permissions_path_set_COLON_mbql.invokeStatic(card.clj:45)" "models.card$permissions_path_set_COLON_mbql.invoke(card.clj:43)" "models.card$query_perms_set_STAR_.invokeStatic(card.clj:67)" "models.card$query_perms_set_STAR_.invoke(card.clj:63)" "api.card$fn__23193$fn__23196.invoke(card.clj:186)" "api.common.internal$do_with_caught_api_exceptions.invokeStatic(internal.clj:222)" "api.common.internal$do_with_caught_api_exceptions.invoke(internal.clj:219)" "api.card$fn__23193.invokeStatic(card.clj:177)" "api.card$fn__23193.invoke(card.clj:177)" "middleware$enforce_authentication$fn__31489.invoke(middleware.clj:84)" "api.routes$fn__31589.invokeStatic(routes.clj:41)" "api.routes$fn__31589.invoke(routes.clj:41)" "routes$fn__32587.invokeStatic(routes.clj:24)" "routes$fn__32587.invoke(routes.clj:24)" "middleware$log_api_call$fn__31577$fn__31579.invoke(middleware.clj:275)" "db$_do_with_call_counting.invokeStatic(db.clj:461)" "db$_do_with_call_counting.invoke(db.clj:455)" "middleware$log_api_call$fn__31577.invoke(middleware.clj:274)" "middleware$add_security_headers$fn__31536.invoke(middleware.clj:209)" "middleware$bind_current_user$fn__31493.invoke(middleware.clj:104)"] 01-13 10:21:23 WARN metabase.middleware :: POST /api/card 403 (117 ms) (2 DB calls) "You don't have permissions to do that." ```
1.0
Non admin users can't see Questions that are using GA as data source - ### Bug Regular User can't access Questions and Cards using data from GoogleAnalytics database but they can ask a question and get data as long as they are not saving it. - Browser and the version: Chrome 55.0.2883.95 - Operating system: macOS Sierra - Databases: MongoDB, GoogleAnalytics - Metabase version: 0.22 - Metabase hosting environment: Docker - Metabase internal database: (default) #### Steps to reproduce - User is not admin - Create new question using Google Analytics as DB source - Get answer - Save the question - Get the permission error : <img width="672" alt="capture d ecran 2017-01-13 a 11 13 59" src="https://cloud.githubusercontent.com/assets/60847/21926465/183da5e8-d982-11e6-8ec4-7808b4c7da09.png"> This is not affecting admin users #### Logs ``` 01-13 10:21:23 WARN models.card :: Error getting permissions for card: Illegal clause (no matching fn found): :metric ["query_processor.expand$fn_for_token.invokeStatic(expand.clj:442)" "query_processor.expand$fn_for_token.invoke(expand.clj:435)" "query_processor.expand$fn__21383$expand_ql_sexpr__21388$fn__21392.invoke(expand.clj:452)" "query_processor.expand$fn__21383$expand_ql_sexpr__21388.invoke(expand.clj:444)" "query_processor.expand$walk_expand_ql_sexprs.invokeStatic(expand.clj:460)" "query_processor.expand$walk_expand_ql_sexprs.invoke(expand.clj:455)" "query_processor.expand$walk_expand_ql_sexprs.invokeStatic(expand.clj:460)" "query_processor.expand$walk_expand_ql_sexprs.invoke(expand.clj:455)" "query_processor.expand$fn__21428$expand_inner__21433$fn__21434.invoke(expand.clj:468)" "query_processor.expand$fn__21428$expand_inner__21433.invoke(expand.clj:464)" "query_processor.expand$expand.invokeStatic(expand.clj:493)" "query_processor.expand$expand.invoke(expand.clj:479)" "query_processor$expand_resolve.invokeStatic(query_processor.clj:178)" "query_processor$expand_resolve.invoke(query_processor.clj:171)" "models.card$permissions_path_set_COLON_mbql.invokeStatic(card.clj:45)" "models.card$permissions_path_set_COLON_mbql.invoke(card.clj:43)" "models.card$query_perms_set_STAR_.invokeStatic(card.clj:67)" "models.card$query_perms_set_STAR_.invoke(card.clj:63)" "api.card$fn__23193$fn__23196.invoke(card.clj:186)" "api.common.internal$do_with_caught_api_exceptions.invokeStatic(internal.clj:222)" "api.common.internal$do_with_caught_api_exceptions.invoke(internal.clj:219)" "api.card$fn__23193.invokeStatic(card.clj:177)" "api.card$fn__23193.invoke(card.clj:177)" "middleware$enforce_authentication$fn__31489.invoke(middleware.clj:84)" "api.routes$fn__31589.invokeStatic(routes.clj:41)" "api.routes$fn__31589.invoke(routes.clj:41)" "routes$fn__32587.invokeStatic(routes.clj:24)" "routes$fn__32587.invoke(routes.clj:24)" "middleware$log_api_call$fn__31577$fn__31579.invoke(middleware.clj:275)" "db$_do_with_call_counting.invokeStatic(db.clj:461)" "db$_do_with_call_counting.invoke(db.clj:455)" "middleware$log_api_call$fn__31577.invoke(middleware.clj:274)" "middleware$add_security_headers$fn__31536.invoke(middleware.clj:209)" "middleware$bind_current_user$fn__31493.invoke(middleware.clj:104)"] 01-13 10:21:23 WARN metabase.middleware :: POST /api/card 403 (117 ms) (2 DB calls) "You don't have permissions to do that." ```
process
non admin users can t see questions that are using ga as data source bug regular user can t access questions and cards using data from googleanalytics database but they can ask a question and get data as long as they are not saving it browser and the version chrome operating system macos sierra databases mongodb googleanalytics metabase version metabase hosting environment docker metabase internal database default steps to reproduce user is not admin create new question using google analytics as db source get answer save the question get the permission error img width alt capture d ecran a src this is not affecting admin users logs warn models card error getting permissions for card illegal clause no matching fn found metric query processor expand fn for token invokestatic expand clj query processor expand fn for token invoke expand clj query processor expand fn expand ql sexpr fn invoke expand clj query processor expand fn expand ql sexpr invoke expand clj query processor expand walk expand ql sexprs invokestatic expand clj query processor expand walk expand ql sexprs invoke expand clj query processor expand walk expand ql sexprs invokestatic expand clj query processor expand walk expand ql sexprs invoke expand clj query processor expand fn expand inner fn invoke expand clj query processor expand fn expand inner invoke expand clj query processor expand expand invokestatic expand clj query processor expand expand invoke expand clj query processor expand resolve invokestatic query processor clj query processor expand resolve invoke query processor clj models card permissions path set colon mbql invokestatic card clj models card permissions path set colon mbql invoke card clj models card query perms set star invokestatic card clj models card query perms set star invoke card clj api card fn fn invoke card clj api common internal do with caught api exceptions invokestatic internal clj api common internal do with caught api exceptions invoke internal clj api card fn invokestatic card clj api card fn invoke card clj middleware enforce authentication fn invoke middleware clj api routes fn invokestatic routes clj api routes fn invoke routes clj routes fn invokestatic routes clj routes fn invoke routes clj middleware log api call fn fn invoke middleware clj db do with call counting invokestatic db clj db do with call counting invoke db clj middleware log api call fn invoke middleware clj middleware add security headers fn invoke middleware clj middleware bind current user fn invoke middleware clj warn metabase middleware post api card ms db calls you don t have permissions to do that
1
20,196
26,772,809,498
IssuesEvent
2023-01-31 15:11:56
Open-EO/openeo-processes
https://api.github.com/repos/Open-EO/openeo-processes
closed
Cropping cumbersome
new process
Feedback from METER group is that cropping to a GeoJSON is cumbersome as you need to run two processes: - [filter_spatial](https://processes.openeo.org/#filter_spatial) and - [mask_polygon](https://processes.openeo.org/#mask_polygon) Might be a client improvement, but maybe we can also enforce implicit filtering in mask_polygon or implicit masking in filter_spatial? Not sure yet...
1.0
Cropping cumbersome - Feedback from METER group is that cropping to a GeoJSON is cumbersome as you need to run two processes: - [filter_spatial](https://processes.openeo.org/#filter_spatial) and - [mask_polygon](https://processes.openeo.org/#mask_polygon) Might be a client improvement, but maybe we can also enforce implicit filtering in mask_polygon or implicit masking in filter_spatial? Not sure yet...
process
cropping cumbersome feedback from meter group is that cropping to a geojson is cumbersome as you need to run two processes and might be a client improvement but maybe we can also enforce implicit filtering in mask polygon or implicit masking in filter spatial not sure yet
1
18,072
24,089,078,497
IssuesEvent
2022-09-19 13:26:59
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
NTR pore forming activity
multi-species process
For the work on multiorganisms, to (re)annotate cytolysis terms, I will create a new MF, "pore forming activity", definition: An activity in which a protein is inserted into the membrane of another cell where it forms transmembrane pores. Added: Pores disrupts the integrity of the cell membrane, resulting in deregulated ion homeostasis, cellular dysfunction, and can result in cell death. TC:1.C PMID:34387717
1.0
NTR pore forming activity - For the work on multiorganisms, to (re)annotate cytolysis terms, I will create a new MF, "pore forming activity", definition: An activity in which a protein is inserted into the membrane of another cell where it forms transmembrane pores. Added: Pores disrupts the integrity of the cell membrane, resulting in deregulated ion homeostasis, cellular dysfunction, and can result in cell death. TC:1.C PMID:34387717
process
ntr pore forming activity for the work on multiorganisms to re annotate cytolysis terms i will create a new mf pore forming activity definition an activity in which a protein is inserted into the membrane of another cell where it forms transmembrane pores added pores disrupts the integrity of the cell membrane resulting in deregulated ion homeostasis cellular dysfunction and can result in cell death tc c pmid
1
696,004
23,879,200,542
IssuesEvent
2022-09-07 22:34:26
oncokb/oncokb
https://api.github.com/repos/oncokb/oncokb
closed
Support cancer type exclusion
high priority Epic next-minor-release
In the recent BRAF/V600E/All Solid Tumors FDA approval, we need to exclude colorectal cancer per approval.
1.0
Support cancer type exclusion - In the recent BRAF/V600E/All Solid Tumors FDA approval, we need to exclude colorectal cancer per approval.
non_process
support cancer type exclusion in the recent braf all solid tumors fda approval we need to exclude colorectal cancer per approval
0
21,969
30,464,224,117
IssuesEvent
2023-07-17 09:13:27
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Invalid environment specification in pipeline YAML targeting resource
doc-bug Pri1 azure-devops-pipelines/svc azure-devops-pipelines-process/subsvc
I believe the YAML in the "Target an environment from a deployment job" section is invalid. When I copy the code block from this section into the YAML pipeline editor, it reports a bunch of errors starting with the `resourceName` line. If I simple move `smarthotel-dev` onto it's own line with the `name` tag, then the errors are resolved. Thus, I think the environment section of the code block should be ```yaml environment: name: 'smarthotel-dev' resourceName: myVM resourceType: virtualMachine ``` This correction is based on the deployment job schema documented here: https://learn.microsoft.com/en-us/azure/devops/pipelines/process/deployment-jobs?view=azure-devops#schema --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Create target environment - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops#target-an-environment-from-a-deployment-job) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments.md) * Service: **azure-devops-pipelines** * Sub-service: **azure-devops-pipelines-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Invalid environment specification in pipeline YAML targeting resource - I believe the YAML in the "Target an environment from a deployment job" section is invalid. When I copy the code block from this section into the YAML pipeline editor, it reports a bunch of errors starting with the `resourceName` line. If I simple move `smarthotel-dev` onto it's own line with the `name` tag, then the errors are resolved. Thus, I think the environment section of the code block should be ```yaml environment: name: 'smarthotel-dev' resourceName: myVM resourceType: virtualMachine ``` This correction is based on the deployment job schema documented here: https://learn.microsoft.com/en-us/azure/devops/pipelines/process/deployment-jobs?view=azure-devops#schema --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Create target environment - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops#target-an-environment-from-a-deployment-job) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments.md) * Service: **azure-devops-pipelines** * Sub-service: **azure-devops-pipelines-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
invalid environment specification in pipeline yaml targeting resource i believe the yaml in the target an environment from a deployment job section is invalid when i copy the code block from this section into the yaml pipeline editor it reports a bunch of errors starting with the resourcename line if i simple move smarthotel dev onto it s own line with the name tag then the errors are resolved thus i think the environment section of the code block should be yaml environment name smarthotel dev resourcename myvm resourcetype virtualmachine this correction is based on the deployment job schema documented here document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service azure devops pipelines sub service azure devops pipelines process github login juliakm microsoft alias jukullam
1
271,363
8,483,195,071
IssuesEvent
2018-10-25 20:48:46
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
closed
Concurrent `update` and `preview` can lead to confusing errors
priority/P1
During a deployment to `pulumi.io` we saw a failure which appears to be related to a concurrent preview and update occuring: https://github.com/pulumi/docs/pull/583#pullrequestreview-154478411 ``` error:the current deployment has 6 resource(s) with pending operations: * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/scheduler/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/search/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/servicefabric/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/sql/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/storage/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure-serverless/index.html, interrupted while updating These resources are in an unknown state because the Pulumi CLI was interrupted while waiting for changes to these resources to complete. You should confirm whether or not the operations listed completed successfully by checking the state of the appropriate provider. For example, if you are using AWS, you can confirm using the AWS Console. Once you have confirmed the status of the interrupted operations, you can repair your stack using 'pulumi stack export' to export your stack to a file. For each operation that succeeded remove that operation from the "pending_operations" section of the file. Once this is complete use 'pulumi stack import' to import the repaired stack. error:refusing to proceed make: *** [preview] Error 255 ```
1.0
Concurrent `update` and `preview` can lead to confusing errors - During a deployment to `pulumi.io` we saw a failure which appears to be related to a concurrent preview and update occuring: https://github.com/pulumi/docs/pull/583#pullrequestreview-154478411 ``` error:the current deployment has 6 resource(s) with pending operations: * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/scheduler/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/search/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/servicefabric/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/sql/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure/storage/index.html, interrupted while updating * urn:pulumi:pulumi.io-staging::pulumi.io::aws:s3/bucket:Bucket$aws:s3/bucketObject:BucketObject::reference/pkg/nodejs/@pulumi/azure-serverless/index.html, interrupted while updating These resources are in an unknown state because the Pulumi CLI was interrupted while waiting for changes to these resources to complete. You should confirm whether or not the operations listed completed successfully by checking the state of the appropriate provider. For example, if you are using AWS, you can confirm using the AWS Console. Once you have confirmed the status of the interrupted operations, you can repair your stack using 'pulumi stack export' to export your stack to a file. For each operation that succeeded remove that operation from the "pending_operations" section of the file. Once this is complete use 'pulumi stack import' to import the repaired stack. error:refusing to proceed make: *** [preview] Error 255 ```
non_process
concurrent update and preview can lead to confusing errors during a deployment to pulumi io we saw a failure which appears to be related to a concurrent preview and update occuring error the current deployment has resource s with pending operations urn pulumi pulumi io staging pulumi io aws bucket bucket aws bucketobject bucketobject reference pkg nodejs pulumi azure scheduler index html interrupted while updating urn pulumi pulumi io staging pulumi io aws bucket bucket aws bucketobject bucketobject reference pkg nodejs pulumi azure search index html interrupted while updating urn pulumi pulumi io staging pulumi io aws bucket bucket aws bucketobject bucketobject reference pkg nodejs pulumi azure servicefabric index html interrupted while updating urn pulumi pulumi io staging pulumi io aws bucket bucket aws bucketobject bucketobject reference pkg nodejs pulumi azure sql index html interrupted while updating urn pulumi pulumi io staging pulumi io aws bucket bucket aws bucketobject bucketobject reference pkg nodejs pulumi azure storage index html interrupted while updating urn pulumi pulumi io staging pulumi io aws bucket bucket aws bucketobject bucketobject reference pkg nodejs pulumi azure serverless index html interrupted while updating these resources are in an unknown state because the pulumi cli was interrupted while waiting for changes to these resources to complete you should confirm whether or not the operations listed completed successfully by checking the state of the appropriate provider for example if you are using aws you can confirm using the aws console once you have confirmed the status of the interrupted operations you can repair your stack using pulumi stack export to export your stack to a file for each operation that succeeded remove that operation from the pending operations section of the file once this is complete use pulumi stack import to import the repaired stack error refusing to proceed make error
0
185,375
14,351,563,318
IssuesEvent
2020-11-30 01:39:52
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
opened
zuoralabs/zsec-go-examples: goroutine_test.go; 9 LoC
fresh test tiny
Found a possible issue in [zuoralabs/zsec-go-examples](https://www.github.com/zuoralabs/zsec-go-examples) at [goroutine_test.go](https://github.com/zuoralabs/zsec-go-examples/blob/605d1788c03ea7e9d2c7594b16f46137ad6e1025/goroutine_test.go#L243-L251) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/zuoralabs/zsec-go-examples/blob/605d1788c03ea7e9d2c7594b16f46137ad6e1025/goroutine_test.go#L243-L251) <details> <summary>Click here to show the 9 line(s) of Go which triggered the analyzer.</summary> ```go for i := range make([]int, numWorkers) { workers.Add(1) go func(k int) { defer workers.Done() time.Sleep(time.Millisecond * 200) fmt.Printf("part 2: working: %v\n", i) mm2.Store(k, i) }(i) } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable i used in defer or goroutine at line 248 </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 605d1788c03ea7e9d2c7594b16f46137ad6e1025
1.0
zuoralabs/zsec-go-examples: goroutine_test.go; 9 LoC - Found a possible issue in [zuoralabs/zsec-go-examples](https://www.github.com/zuoralabs/zsec-go-examples) at [goroutine_test.go](https://github.com/zuoralabs/zsec-go-examples/blob/605d1788c03ea7e9d2c7594b16f46137ad6e1025/goroutine_test.go#L243-L251) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/zuoralabs/zsec-go-examples/blob/605d1788c03ea7e9d2c7594b16f46137ad6e1025/goroutine_test.go#L243-L251) <details> <summary>Click here to show the 9 line(s) of Go which triggered the analyzer.</summary> ```go for i := range make([]int, numWorkers) { workers.Add(1) go func(k int) { defer workers.Done() time.Sleep(time.Millisecond * 200) fmt.Printf("part 2: working: %v\n", i) mm2.Store(k, i) }(i) } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable i used in defer or goroutine at line 248 </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 605d1788c03ea7e9d2c7594b16f46137ad6e1025
non_process
zuoralabs zsec go examples goroutine test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for i range make int numworkers workers add go func k int defer workers done time sleep time millisecond fmt printf part working v n i store k i i below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable i used in defer or goroutine at line leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
15,963
20,177,192,988
IssuesEvent
2022-02-10 15:24:19
ossf/tac
https://api.github.com/repos/ossf/tac
closed
TAC Election: Officials
ElectionProcess
Who should be the Election Officials? Suggestion: - OpenSSF General Manager - OpenSSF Program Director - OpenSSF TAC Chair - OpenSSF TAC rep from an org that is not the same org as the TAC Chair
1.0
TAC Election: Officials - Who should be the Election Officials? Suggestion: - OpenSSF General Manager - OpenSSF Program Director - OpenSSF TAC Chair - OpenSSF TAC rep from an org that is not the same org as the TAC Chair
process
tac election officials who should be the election officials suggestion openssf general manager openssf program director openssf tac chair openssf tac rep from an org that is not the same org as the tac chair
1
14,097
16,920,313,411
IssuesEvent
2021-06-25 03:58:34
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
UPDATE result is not compatible with MySQL
compatibility-breaker severity/moderate sig/sql-infra status/future type/bug type/wontfix
## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) ``` create table companies(id bigint primary key, ida bigint); insert into companies values(14, 14); UPDATE companies SET id = id + 1, ida = id * 2; ``` <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) in the mysql, you can see ![image](https://user-images.githubusercontent.com/3427324/122853573-05e48a80-d345-11eb-9568-11bd872fa303.png) ### 3. What did you see instead (Required) but in the tidb ![image](https://user-images.githubusercontent.com/3427324/122853658-26144980-d345-11eb-86fe-569a84741fbc.png) ### 4. What is your TiDB version? (Required) Release Version: v5.2.0-alpha-122-g3132b5940-dirty Edition: Community Git Commit Hash: 3132b59402179fe6859dd17f84fec0a7173624b0 Git Branch: master <!-- Paste the output of SELECT tidb_version() -->
True
UPDATE result is not compatible with MySQL - ## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) ``` create table companies(id bigint primary key, ida bigint); insert into companies values(14, 14); UPDATE companies SET id = id + 1, ida = id * 2; ``` <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) in the mysql, you can see ![image](https://user-images.githubusercontent.com/3427324/122853573-05e48a80-d345-11eb-9568-11bd872fa303.png) ### 3. What did you see instead (Required) but in the tidb ![image](https://user-images.githubusercontent.com/3427324/122853658-26144980-d345-11eb-86fe-569a84741fbc.png) ### 4. What is your TiDB version? (Required) Release Version: v5.2.0-alpha-122-g3132b5940-dirty Edition: Community Git Commit Hash: 3132b59402179fe6859dd17f84fec0a7173624b0 Git Branch: master <!-- Paste the output of SELECT tidb_version() -->
non_process
update result is not compatible with mysql bug report please answer these questions before submitting your issue thanks minimal reproduce step required create table companies id bigint primary key ida bigint insert into companies values update companies set id id ida id what did you expect to see required in the mysql you can see what did you see instead required but in the tidb what is your tidb version required release version alpha dirty edition community git commit hash git branch master
0
470,748
13,543,702,414
IssuesEvent
2020-09-16 19:23:51
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Errors in the HL7800.c file
bug priority: low
**Describe the bug** I have a Pinnacle 100 dvk board that I integrated to Platformio to be used with Zephyr 2.3. I need to use the HL7800 semiconductor that is not available in the current version of Zephyr but may be for the 2.4 version. So I decided to add what is missing manually, so far so good. I end up with these errors: **1)** > ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c: In function 'mdm_gpio6_callback_isr': ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c:3607:2: error: #endif without #if #endif ^~~~~ **2)** I tried to fix this one but I am missing something... > ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c: In function 'hl7800_init': ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c:4862:15: error: 'struct mdm_receiver_context' has no member named 'data_imei'; did you mean 'data_iccid'? ictx.mdm_ctx.data_imei = ictx.mdm_imei; ^~~~~~~~~ data_iccid **3)** > At top level: ~.platformio/packages/framework-zephyr/drivers/modem/hl7800.c:322:19: warning: 'TIME_STRING_FORMAT' defined but not used [-Wunused-const-variable=] static const char TIME_STRING_FORMAT[] = "\"yy/MM/dd,hh:mm:ss?zz\""; **Environment (please complete the following information):** - OS: Linux, - Toolchain Zephyr SDK
1.0
Errors in the HL7800.c file - **Describe the bug** I have a Pinnacle 100 dvk board that I integrated to Platformio to be used with Zephyr 2.3. I need to use the HL7800 semiconductor that is not available in the current version of Zephyr but may be for the 2.4 version. So I decided to add what is missing manually, so far so good. I end up with these errors: **1)** > ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c: In function 'mdm_gpio6_callback_isr': ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c:3607:2: error: #endif without #if #endif ^~~~~ **2)** I tried to fix this one but I am missing something... > ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c: In function 'hl7800_init': ~/.platformio/packages/framework-zephyr/drivers/modem/hl7800.c:4862:15: error: 'struct mdm_receiver_context' has no member named 'data_imei'; did you mean 'data_iccid'? ictx.mdm_ctx.data_imei = ictx.mdm_imei; ^~~~~~~~~ data_iccid **3)** > At top level: ~.platformio/packages/framework-zephyr/drivers/modem/hl7800.c:322:19: warning: 'TIME_STRING_FORMAT' defined but not used [-Wunused-const-variable=] static const char TIME_STRING_FORMAT[] = "\"yy/MM/dd,hh:mm:ss?zz\""; **Environment (please complete the following information):** - OS: Linux, - Toolchain Zephyr SDK
non_process
errors in the c file describe the bug i have a pinnacle dvk board that i integrated to platformio to be used with zephyr i need to use the semiconductor that is not available in the current version of zephyr but may be for the version so i decided to add what is missing manually so far so good i end up with these errors platformio packages framework zephyr drivers modem c in function mdm callback isr platformio packages framework zephyr drivers modem c error endif without if endif i tried to fix this one but i am missing something platformio packages framework zephyr drivers modem c in function init platformio packages framework zephyr drivers modem c error struct mdm receiver context has no member named data imei did you mean data iccid ictx mdm ctx data imei ictx mdm imei data iccid at top level platformio packages framework zephyr drivers modem c warning time string format defined but not used static const char time string format yy mm dd hh mm ss zz environment please complete the following information os linux toolchain zephyr sdk
0
195,995
15,570,791,696
IssuesEvent
2021-03-17 03:24:53
gravitational/teleport
https://api.github.com/repos/gravitational/teleport
opened
Fixes for next docs linter
documentation release-engineering
- Backport `test-docs` to at least `branch/v6` so that docs PRs to that branch are linted. - Maybe `branch/5.0` as well if we're planning to make any changes to those docs as a matter of course - Maybe it even needs to be backported to all older branches which run Drone as they're all using the new docs engine now. - Figure out why warnings don't cause Drone to exit the pipeline with errors (https://drone.gravitational.io/gravitational/teleport/7920/2/3 for example)
1.0
Fixes for next docs linter - - Backport `test-docs` to at least `branch/v6` so that docs PRs to that branch are linted. - Maybe `branch/5.0` as well if we're planning to make any changes to those docs as a matter of course - Maybe it even needs to be backported to all older branches which run Drone as they're all using the new docs engine now. - Figure out why warnings don't cause Drone to exit the pipeline with errors (https://drone.gravitational.io/gravitational/teleport/7920/2/3 for example)
non_process
fixes for next docs linter backport test docs to at least branch so that docs prs to that branch are linted maybe branch as well if we re planning to make any changes to those docs as a matter of course maybe it even needs to be backported to all older branches which run drone as they re all using the new docs engine now figure out why warnings don t cause drone to exit the pipeline with errors for example
0
17,722
23,625,566,596
IssuesEvent
2022-08-25 03:17:09
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add The Attack of the Amazing Electrified Man from Popcorn (1991)
suggested title in process
Title: The Attack of the Amazing Electrified Man Type (film/tv show): Film Film or show in which it appears: Popcorn (1991) Is the parent film/show streaming anywhere? Yes (Shudder and AMC+); https://www.justwatch.com/us/movie/popcorn-1991 About when in the parent film/show does it appear? 0:39:00 Actual footage of the film/show can be seen (yes/no)? yes Quotes: - Tell me what do you know about erythrocytes and leukocytes? - Say, aren't those the red and white corpuscles in the blood? - Very good, Vernon. I have created a third corpuscle the "magnecyte." Theoretically, it can regenerate the blood cells even after severe trauma, any trauma. Even electrocution. - I get it. You want to inject me with this corpuscle to see if I can survive the chair, is that it? - Exactly. Screenshots: ![taotaem_1](https://user-images.githubusercontent.com/80119678/183198882-d9ed822b-bc8e-493d-84e5-c880006ed831.png) ![taotaem_2](https://user-images.githubusercontent.com/80119678/183198912-f5e0e048-ae58-4788-9528-d60d81b5da31.png) ![taotaem_3](https://user-images.githubusercontent.com/80119678/183198924-26a826fc-b592-4c18-ae9e-d00af85b574f.png) ![taotaem_4](https://user-images.githubusercontent.com/80119678/183198935-11a9b76b-0148-44d4-9337-3abedd7353b8.png) ![taotaem_5](https://user-images.githubusercontent.com/80119678/183198946-cc3ad23a-806a-4161-9885-5f0a3d7c993e.png) ![taotaem_6](https://user-images.githubusercontent.com/80119678/183198957-59626e71-4f63-4d83-a995-dde9787b8f00.png) ![taotaem_7](https://user-images.githubusercontent.com/80119678/183198968-9cf01a58-4c53-4b10-9ecf-0c736b5fefe1.png) ![taotaem_8](https://user-images.githubusercontent.com/80119678/183198978-c58a3fa6-6d5a-4e6a-8fdb-08a814928bc2.png) ![taotaem_9](https://user-images.githubusercontent.com/80119678/183198989-1f8b2df4-950c-49f4-bbeb-dc645e41d3a7.png) ![taotaem_10](https://user-images.githubusercontent.com/80119678/183198999-1bc7f650-aaea-45f2-bc33-087875426a64.png) ![taotaem_11](https://user-images.githubusercontent.com/80119678/183199012-80f4533a-b3dc-46b6-819a-06b653477068.png) ![taotaem_12](https://user-images.githubusercontent.com/80119678/183199021-0cf2aa41-518a-42b2-aa2c-85c54a3d2bd0.png) ![taotaem_13](https://user-images.githubusercontent.com/80119678/183199031-f235b392-93b8-4d64-a3f3-e44ab62326c9.png) ![taotaem_14](https://user-images.githubusercontent.com/80119678/183199043-466d3eee-10d7-4ce3-a4cb-271e3e241c74.png) ![taotaem_15](https://user-images.githubusercontent.com/80119678/183199050-9c03cd72-7972-4cb3-9e8e-9a9dfe8bd13b.png) ![taotaem_16](https://user-images.githubusercontent.com/80119678/183199066-12d24fb2-2353-483a-8f96-f70b3e0cad21.png)
1.0
Add The Attack of the Amazing Electrified Man from Popcorn (1991) - Title: The Attack of the Amazing Electrified Man Type (film/tv show): Film Film or show in which it appears: Popcorn (1991) Is the parent film/show streaming anywhere? Yes (Shudder and AMC+); https://www.justwatch.com/us/movie/popcorn-1991 About when in the parent film/show does it appear? 0:39:00 Actual footage of the film/show can be seen (yes/no)? yes Quotes: - Tell me what do you know about erythrocytes and leukocytes? - Say, aren't those the red and white corpuscles in the blood? - Very good, Vernon. I have created a third corpuscle the "magnecyte." Theoretically, it can regenerate the blood cells even after severe trauma, any trauma. Even electrocution. - I get it. You want to inject me with this corpuscle to see if I can survive the chair, is that it? - Exactly. Screenshots: ![taotaem_1](https://user-images.githubusercontent.com/80119678/183198882-d9ed822b-bc8e-493d-84e5-c880006ed831.png) ![taotaem_2](https://user-images.githubusercontent.com/80119678/183198912-f5e0e048-ae58-4788-9528-d60d81b5da31.png) ![taotaem_3](https://user-images.githubusercontent.com/80119678/183198924-26a826fc-b592-4c18-ae9e-d00af85b574f.png) ![taotaem_4](https://user-images.githubusercontent.com/80119678/183198935-11a9b76b-0148-44d4-9337-3abedd7353b8.png) ![taotaem_5](https://user-images.githubusercontent.com/80119678/183198946-cc3ad23a-806a-4161-9885-5f0a3d7c993e.png) ![taotaem_6](https://user-images.githubusercontent.com/80119678/183198957-59626e71-4f63-4d83-a995-dde9787b8f00.png) ![taotaem_7](https://user-images.githubusercontent.com/80119678/183198968-9cf01a58-4c53-4b10-9ecf-0c736b5fefe1.png) ![taotaem_8](https://user-images.githubusercontent.com/80119678/183198978-c58a3fa6-6d5a-4e6a-8fdb-08a814928bc2.png) ![taotaem_9](https://user-images.githubusercontent.com/80119678/183198989-1f8b2df4-950c-49f4-bbeb-dc645e41d3a7.png) ![taotaem_10](https://user-images.githubusercontent.com/80119678/183198999-1bc7f650-aaea-45f2-bc33-087875426a64.png) ![taotaem_11](https://user-images.githubusercontent.com/80119678/183199012-80f4533a-b3dc-46b6-819a-06b653477068.png) ![taotaem_12](https://user-images.githubusercontent.com/80119678/183199021-0cf2aa41-518a-42b2-aa2c-85c54a3d2bd0.png) ![taotaem_13](https://user-images.githubusercontent.com/80119678/183199031-f235b392-93b8-4d64-a3f3-e44ab62326c9.png) ![taotaem_14](https://user-images.githubusercontent.com/80119678/183199043-466d3eee-10d7-4ce3-a4cb-271e3e241c74.png) ![taotaem_15](https://user-images.githubusercontent.com/80119678/183199050-9c03cd72-7972-4cb3-9e8e-9a9dfe8bd13b.png) ![taotaem_16](https://user-images.githubusercontent.com/80119678/183199066-12d24fb2-2353-483a-8f96-f70b3e0cad21.png)
process
add the attack of the amazing electrified man from popcorn title the attack of the amazing electrified man type film tv show film film or show in which it appears popcorn is the parent film show streaming anywhere yes shudder and amc about when in the parent film show does it appear actual footage of the film show can be seen yes no yes quotes tell me what do you know about erythrocytes and leukocytes say aren t those the red and white corpuscles in the blood very good vernon i have created a third corpuscle the magnecyte theoretically it can regenerate the blood cells even after severe trauma any trauma even electrocution i get it you want to inject me with this corpuscle to see if i can survive the chair is that it exactly screenshots
1
4,440
3,024,885,908
IssuesEvent
2015-08-03 01:55:11
catapult-project/catapult
https://api.github.com/repos/catapult-project/catapult
opened
Clarify role of IR type, IR title using an override of userFriendlyName
Code Health
<a href="https://github.com/natduca"><img src="https://avatars.githubusercontent.com/u/412396?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [natduca](https://github.com/natduca)** _Thursday Jul 16, 2015 at 22:52 GMT_ _Originally opened as https://github.com/google/trace-viewer/issues/1099_ ---- IR.displayName returns type + ': ' + title title is what the user is doing, e.g. click, drag, keytyping ir type is R/A/I/L types (upper)
1.0
Clarify role of IR type, IR title using an override of userFriendlyName - <a href="https://github.com/natduca"><img src="https://avatars.githubusercontent.com/u/412396?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [natduca](https://github.com/natduca)** _Thursday Jul 16, 2015 at 22:52 GMT_ _Originally opened as https://github.com/google/trace-viewer/issues/1099_ ---- IR.displayName returns type + ': ' + title title is what the user is doing, e.g. click, drag, keytyping ir type is R/A/I/L types (upper)
non_process
clarify role of ir type ir title using an override of userfriendlyname issue by thursday jul at gmt originally opened as ir displayname returns type title title is what the user is doing e g click drag keytyping ir type is r a i l types upper
0
42,276
2,869,997,020
IssuesEvent
2015-06-06 18:37:38
thexerteproject/xerteonlinetoolkits
https://api.github.com/repos/thexerteproject/xerteonlinetoolkits
opened
Turning on DB authentication causes error loading /index.php
bug high priority
You get this error Strict Standards: Declaration of Xerte_Authentication_Db::addUser() should be compatible with Xerte_Authentication_Abstract::addUser($username, $passwd, $firstname, $lastname) in /home/domain/library/Xerte/Authentication/Db.php on line 35
1.0
Turning on DB authentication causes error loading /index.php - You get this error Strict Standards: Declaration of Xerte_Authentication_Db::addUser() should be compatible with Xerte_Authentication_Abstract::addUser($username, $passwd, $firstname, $lastname) in /home/domain/library/Xerte/Authentication/Db.php on line 35
non_process
turning on db authentication causes error loading index php you get this error strict standards declaration of xerte authentication db adduser should be compatible with xerte authentication abstract adduser username passwd firstname lastname in home domain library xerte authentication db php on line
0
8,381
11,543,404,307
IssuesEvent
2020-02-18 09:32:39
google/go-jsonnet
https://api.github.com/repos/google/go-jsonnet
closed
Add tests for 32-bit to Travis
process
As far as I can tell, there is no i686 arch in Travis available, but we can just compile stuff in 32-bit mode. It should let us catch the problems like #376 early.
1.0
Add tests for 32-bit to Travis - As far as I can tell, there is no i686 arch in Travis available, but we can just compile stuff in 32-bit mode. It should let us catch the problems like #376 early.
process
add tests for bit to travis as far as i can tell there is no arch in travis available but we can just compile stuff in bit mode it should let us catch the problems like early
1
726,490
25,000,784,634
IssuesEvent
2022-11-03 07:40:39
submariner-io/enhancements
https://api.github.com/repos/submariner-io/enhancements
closed
Epic: Support OCP with submariner on VMWare
enhancement priority:high
**What would you like to be added**: Proposal is to fully support connecting multiple OCP clusters on VMWare using Submariner. **Why is this needed**: Aim of this proposal is to list and track steps needed to fill any gaps and allow full support for such deployments. **UPDATE** Investigations have confirmed that there is nothing that needs to be done to prepare cloud for VM Ware, in subctl or OCM. Expectation is that VMWare should work without any changes, except when using NSX which is for now outside scope of this EPIC. **Work Items** Based on results of investigations, only effort required is testing. - [ ] Add VMWare to Submariner Testday - [ ] Run QE on VMWare
1.0
Epic: Support OCP with submariner on VMWare - **What would you like to be added**: Proposal is to fully support connecting multiple OCP clusters on VMWare using Submariner. **Why is this needed**: Aim of this proposal is to list and track steps needed to fill any gaps and allow full support for such deployments. **UPDATE** Investigations have confirmed that there is nothing that needs to be done to prepare cloud for VM Ware, in subctl or OCM. Expectation is that VMWare should work without any changes, except when using NSX which is for now outside scope of this EPIC. **Work Items** Based on results of investigations, only effort required is testing. - [ ] Add VMWare to Submariner Testday - [ ] Run QE on VMWare
non_process
epic support ocp with submariner on vmware what would you like to be added proposal is to fully support connecting multiple ocp clusters on vmware using submariner why is this needed aim of this proposal is to list and track steps needed to fill any gaps and allow full support for such deployments update investigations have confirmed that there is nothing that needs to be done to prepare cloud for vm ware in subctl or ocm expectation is that vmware should work without any changes except when using nsx which is for now outside scope of this epic work items based on results of investigations only effort required is testing add vmware to submariner testday run qe on vmware
0
99,955
8,719,190,487
IssuesEvent
2018-12-07 23:21:08
OpenLiberty/open-liberty
https://api.github.com/repos/OpenLiberty/open-liberty
closed
Update com.ibm.ws.microprofile.metrics_fat_tck dependencies
team:Lumberjack test bug
com.ibm.ws.microprofile.metrics_fat_tck was still using an internal dependency on Arquillian and external contributors couldn't run it.
1.0
Update com.ibm.ws.microprofile.metrics_fat_tck dependencies - com.ibm.ws.microprofile.metrics_fat_tck was still using an internal dependency on Arquillian and external contributors couldn't run it.
non_process
update com ibm ws microprofile metrics fat tck dependencies com ibm ws microprofile metrics fat tck was still using an internal dependency on arquillian and external contributors couldn t run it
0
271
5,315,392,461
IssuesEvent
2017-02-13 17:12:18
PopulateTools/gobierto
https://api.github.com/repos/PopulateTools/gobierto
closed
Mark days with events in the calendar
enhancement gobierto-people
The days that have events should be higlighted and differentiated from the days without events.
1.0
Mark days with events in the calendar - The days that have events should be higlighted and differentiated from the days without events.
non_process
mark days with events in the calendar the days that have events should be higlighted and differentiated from the days without events
0
367
6,225,170,496
IssuesEvent
2017-07-10 15:38:58
offa/danek
https://api.github.com/repos/offa/danek
closed
Refactor BufferedFileReader class
enhancement portability
Refactor `BufferedFileReader` class. - [x] Use RAII - [x] Disable copying - [x] Clean up the special member functions - [x] Proper portability - [x] Use better error handling - [x] Remove preprocessor usage - [x] Avoid C API's as much as possible - [x] Use better types whenever possible - [x] Move to a dedicated header
True
Refactor BufferedFileReader class - Refactor `BufferedFileReader` class. - [x] Use RAII - [x] Disable copying - [x] Clean up the special member functions - [x] Proper portability - [x] Use better error handling - [x] Remove preprocessor usage - [x] Avoid C API's as much as possible - [x] Use better types whenever possible - [x] Move to a dedicated header
non_process
refactor bufferedfilereader class refactor bufferedfilereader class use raii disable copying clean up the special member functions proper portability use better error handling remove preprocessor usage avoid c api s as much as possible use better types whenever possible move to a dedicated header
0
315,569
27,085,615,951
IssuesEvent
2023-02-14 16:47:05
arfc/openmcyclus
https://api.github.com/repos/arfc/openmcyclus
closed
Need CI in repo
Comp:Core Difficulty:1-Beginner Priority:2-Normal Status:1-New Type:Test
This repository needs CI implemented to run tests. This issue can be closed when: - [ ] CI is set up - [ ] CI is called when a new PR is created. I recommend using GitHub actions to keep things contained in this repository instead of calling a third-party.
1.0
Need CI in repo - This repository needs CI implemented to run tests. This issue can be closed when: - [ ] CI is set up - [ ] CI is called when a new PR is created. I recommend using GitHub actions to keep things contained in this repository instead of calling a third-party.
non_process
need ci in repo this repository needs ci implemented to run tests this issue can be closed when ci is set up ci is called when a new pr is created i recommend using github actions to keep things contained in this repository instead of calling a third party
0
295,845
9,101,457,720
IssuesEvent
2019-02-20 11:08:52
jetstack/cert-manager
https://api.github.com/repos/jetstack/cert-manager
closed
Validation stops certificate creation
area/api kind/bug priority/important-soon
While trying to use cert-manager 0.6.1 with the no-webhook configuration, on a GKE cluster running 1.11.6, I encountered this error in the cert-manager logs `status.certificate in body must be of type string: "null"`. Context with minor redaction...: ``` cert-manager-7476cc944f-7m9cg cert-manager I0216 04:11:06.017330 1 controller.go:145] certificates controller: syncing item 'artifactserver/wildcard-example-com' cert-manager-7476cc944f-7m9cg cert-manager I0216 04:11:06.024276 1 sync.go:354] Error issuing certificate for artifactserver/wildcard-example-com: Order.certmanager.k8s.io "wildcard-example-com-1312178538" is invalid: []: Invalid value: map[string]interface {}{"kind":"Order", "apiVersion":"certmanager.k8s.io/v1alpha1", "metadata":map[string]interface {}{"uid":"e18e1cad-31a0-11e9-a817-42010a80006c", "selfLink":"", "name":"wildcard-example-com-1312178538", "namespace":"artifactserver", "creationTimestamp":"2019-02-16T04:11:06Z", "labels":map[string]interface {}{"acme.cert-manager.io/certificate-name":"wildcard-example-com"}, "ownerReferences":[]interface {}{map[string]interface {}{"name":"wildcard-example-com", "uid":"5c1d5dac-31a0-11e9-a817-42010a80006c", "controller":true, "blockOwnerDeletion":true, "apiVersion":"certmanager.k8s.io/v1alpha1", "kind":"Certificate"}}, "generation":1}, "spec":map[string]interface {}{"dnsNames":[]interface {}{"*.example.com"}, "config":[]interface {}{map[string]interface {}{"domains":[]interface {}{"*.example.com"}, "dns01":map[string]interface {}{"provider":"default"}}}, "csr":"...", "issuerRef":map[string]interface {}{"name":"letsencrypt"}}, "status":map[string]interface {}{"state":"", "reason":"", "url":"", "finalizeURL":"", "certificate":interface {}(nil)}}: validation failure list: cert-manager-7476cc944f-7m9cg cert-manager status.certificate in body must be of type string: "null" ``` By removing the validation block from the Order CRD entirely I was able to proceed, and a certificate was issued (yay). But it looks like status.certificate is being treated as a required field.
1.0
Validation stops certificate creation - While trying to use cert-manager 0.6.1 with the no-webhook configuration, on a GKE cluster running 1.11.6, I encountered this error in the cert-manager logs `status.certificate in body must be of type string: "null"`. Context with minor redaction...: ``` cert-manager-7476cc944f-7m9cg cert-manager I0216 04:11:06.017330 1 controller.go:145] certificates controller: syncing item 'artifactserver/wildcard-example-com' cert-manager-7476cc944f-7m9cg cert-manager I0216 04:11:06.024276 1 sync.go:354] Error issuing certificate for artifactserver/wildcard-example-com: Order.certmanager.k8s.io "wildcard-example-com-1312178538" is invalid: []: Invalid value: map[string]interface {}{"kind":"Order", "apiVersion":"certmanager.k8s.io/v1alpha1", "metadata":map[string]interface {}{"uid":"e18e1cad-31a0-11e9-a817-42010a80006c", "selfLink":"", "name":"wildcard-example-com-1312178538", "namespace":"artifactserver", "creationTimestamp":"2019-02-16T04:11:06Z", "labels":map[string]interface {}{"acme.cert-manager.io/certificate-name":"wildcard-example-com"}, "ownerReferences":[]interface {}{map[string]interface {}{"name":"wildcard-example-com", "uid":"5c1d5dac-31a0-11e9-a817-42010a80006c", "controller":true, "blockOwnerDeletion":true, "apiVersion":"certmanager.k8s.io/v1alpha1", "kind":"Certificate"}}, "generation":1}, "spec":map[string]interface {}{"dnsNames":[]interface {}{"*.example.com"}, "config":[]interface {}{map[string]interface {}{"domains":[]interface {}{"*.example.com"}, "dns01":map[string]interface {}{"provider":"default"}}}, "csr":"...", "issuerRef":map[string]interface {}{"name":"letsencrypt"}}, "status":map[string]interface {}{"state":"", "reason":"", "url":"", "finalizeURL":"", "certificate":interface {}(nil)}}: validation failure list: cert-manager-7476cc944f-7m9cg cert-manager status.certificate in body must be of type string: "null" ``` By removing the validation block from the Order CRD entirely I was able to proceed, and a certificate was issued (yay). But it looks like status.certificate is being treated as a required field.
non_process
validation stops certificate creation while trying to use cert manager with the no webhook configuration on a gke cluster running i encountered this error in the cert manager logs status certificate in body must be of type string null context with minor redaction cert manager cert manager controller go certificates controller syncing item artifactserver wildcard example com cert manager cert manager sync go error issuing certificate for artifactserver wildcard example com order certmanager io wildcard example com is invalid invalid value map interface kind order apiversion certmanager io metadata map interface uid selflink name wildcard example com namespace artifactserver creationtimestamp labels map interface acme cert manager io certificate name wildcard example com ownerreferences interface map interface name wildcard example com uid controller true blockownerdeletion true apiversion certmanager io kind certificate generation spec map interface dnsnames interface example com config interface map interface domains interface example com map interface provider default csr issuerref map interface name letsencrypt status map interface state reason url finalizeurl certificate interface nil validation failure list cert manager cert manager status certificate in body must be of type string null by removing the validation block from the order crd entirely i was able to proceed and a certificate was issued yay but it looks like status certificate is being treated as a required field
0
8,389
11,562,704,650
IssuesEvent
2020-02-20 03:25:53
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Obsolete negative regulation by symbiont of entry into host and children
multi-species process
Following up on #18778. The symbiont doesn't negatively regulates its own entry into the host. Will obsolete * GO:0052374 negative regulation by symbiont of entry into host * GO:0075056 negative regulation of symbiont penetration peg formation for entry into host * GO:0075200 negative regulation of symbiont haustorium neck formation for entry into host * GO:0075204 negative regulation of symbiont penetration hypha formation for entry into host No annotations, no mappings, not in any subsets. Thanks, Pascale
1.0
Obsolete negative regulation by symbiont of entry into host and children - Following up on #18778. The symbiont doesn't negatively regulates its own entry into the host. Will obsolete * GO:0052374 negative regulation by symbiont of entry into host * GO:0075056 negative regulation of symbiont penetration peg formation for entry into host * GO:0075200 negative regulation of symbiont haustorium neck formation for entry into host * GO:0075204 negative regulation of symbiont penetration hypha formation for entry into host No annotations, no mappings, not in any subsets. Thanks, Pascale
process
obsolete negative regulation by symbiont of entry into host and children following up on the symbiont doesn t negatively regulates its own entry into the host will obsolete go negative regulation by symbiont of entry into host go negative regulation of symbiont penetration peg formation for entry into host go negative regulation of symbiont haustorium neck formation for entry into host go negative regulation of symbiont penetration hypha formation for entry into host no annotations no mappings not in any subsets thanks pascale
1
269,940
8,444,657,077
IssuesEvent
2018-10-18 19:05:06
DoSomething/blink
https://api.github.com/repos/DoSomething/blink
closed
RabbitMQ uses about 70% of the memory available.
Maintenance Priority (High)
The memory utilization reported by the node in the RabbitMQ interface averages about 70%. We should research what is making it consume so much memory and if it's a signal that we should scale and/or optimize anything.
1.0
RabbitMQ uses about 70% of the memory available. - The memory utilization reported by the node in the RabbitMQ interface averages about 70%. We should research what is making it consume so much memory and if it's a signal that we should scale and/or optimize anything.
non_process
rabbitmq uses about of the memory available the memory utilization reported by the node in the rabbitmq interface averages about we should research what is making it consume so much memory and if it s a signal that we should scale and or optimize anything
0
818,771
30,703,317,107
IssuesEvent
2023-07-27 02:41:47
Memmy-App/memmy
https://api.github.com/repos/Memmy-App/memmy
closed
Apply „Display total score“ option everywhere
enhancement medium priority visual help wanted
**Is your feature request related to a problem? Please describe.** Currently, the option to display the total score instead of up/downvotes only applies to the posts on standard feed view. **Describe the solution you'd like** Have the option also apply to other instances where we see up and downvotes, like posts in compact feed view and comments **Describe alternatives you've considered** / **Additional context** /
1.0
Apply „Display total score“ option everywhere - **Is your feature request related to a problem? Please describe.** Currently, the option to display the total score instead of up/downvotes only applies to the posts on standard feed view. **Describe the solution you'd like** Have the option also apply to other instances where we see up and downvotes, like posts in compact feed view and comments **Describe alternatives you've considered** / **Additional context** /
non_process
apply „display total score“ option everywhere is your feature request related to a problem please describe currently the option to display the total score instead of up downvotes only applies to the posts on standard feed view describe the solution you d like have the option also apply to other instances where we see up and downvotes like posts in compact feed view and comments describe alternatives you ve considered additional context
0
13,779
16,537,543,973
IssuesEvent
2021-05-27 13:27:20
encode/uvicorn
https://api.github.com/repos/encode/uvicorn
closed
Hot-reloading gets stuck after interacting with multiprocessing / Twisted signals
multiprocessing need confirmation workers
### Checklist <!-- Please make sure you check all these items before submitting your bug report. --> - [x] The bug is reproducible against the latest release and/or `master`. - [x] There are no similar issues or pull requests to fix it yet. ### Describe the bug Launched with `--reload` option gets stuck after executing code that interacts with multiprocessing. More details below. ### To reproduce I prepared a minimal build to reproduce this error: https://github.com/desprit/uvicorn_reload_test ### Expected behavior Uvicorn restarts. ### Actual behavior Uvicorn get's stuck with message `WARNING: Detected file change in...` ### Environment `v0.12.2` `uvicorn main:app --reload` ### Additional context I tested it with [hypercorn](https://pgjones.gitlab.io/hypercorn/) and hot reloading doesn't break there. Did also tests with Flask + Gunicorn, works well there as well.
1.0
Hot-reloading gets stuck after interacting with multiprocessing / Twisted signals - ### Checklist <!-- Please make sure you check all these items before submitting your bug report. --> - [x] The bug is reproducible against the latest release and/or `master`. - [x] There are no similar issues or pull requests to fix it yet. ### Describe the bug Launched with `--reload` option gets stuck after executing code that interacts with multiprocessing. More details below. ### To reproduce I prepared a minimal build to reproduce this error: https://github.com/desprit/uvicorn_reload_test ### Expected behavior Uvicorn restarts. ### Actual behavior Uvicorn get's stuck with message `WARNING: Detected file change in...` ### Environment `v0.12.2` `uvicorn main:app --reload` ### Additional context I tested it with [hypercorn](https://pgjones.gitlab.io/hypercorn/) and hot reloading doesn't break there. Did also tests with Flask + Gunicorn, works well there as well.
process
hot reloading gets stuck after interacting with multiprocessing twisted signals checklist the bug is reproducible against the latest release and or master there are no similar issues or pull requests to fix it yet describe the bug launched with reload option gets stuck after executing code that interacts with multiprocessing more details below to reproduce i prepared a minimal build to reproduce this error expected behavior uvicorn restarts actual behavior uvicorn get s stuck with message warning detected file change in environment uvicorn main app reload additional context i tested it with and hot reloading doesn t break there did also tests with flask gunicorn works well there as well
1
14,280
17,260,316,565
IssuesEvent
2021-07-22 06:29:33
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Open study > Enrollment registry > Enrollment target > Text case issue
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
Open study > Enrollment registry > Click on enrollment target > 'Edit Enrollment Target' should be changed to 'Edit enrollment target' ![et1](https://user-images.githubusercontent.com/71445210/125257237-d65ce880-e31a-11eb-966b-851af2a9e0d1.png)
3.0
[PM] Open study > Enrollment registry > Enrollment target > Text case issue - Open study > Enrollment registry > Click on enrollment target > 'Edit Enrollment Target' should be changed to 'Edit enrollment target' ![et1](https://user-images.githubusercontent.com/71445210/125257237-d65ce880-e31a-11eb-966b-851af2a9e0d1.png)
process
open study enrollment registry enrollment target text case issue open study enrollment registry click on enrollment target edit enrollment target should be changed to edit enrollment target
1
180,945
30,594,225,997
IssuesEvent
2023-07-21 20:05:30
ni/nimble
https://api.github.com/repos/ni/nimble
opened
Allow user to remove horizontal scroll space introduced by sizing table columns
enhancement triage IxD visual design
## 😯 Problem to Solve The current behavior for interactive column resizing allows for the creation of a horizontal scrollable area. At the moment, once created, the scrollable area can not be removed sans refreshing the page, which isn't great. It would be nice for us to provide the user with a means of removing the extra horizontal scrollable space. ## 💁 Proposed Solution There has already been some visual/interactive design for this problem. Some considered solutions include: - Providing a button that appears once there is a horizontal scrollbar present that, when pressed, removes any excess horizontal space, while maintaining existing column proportional sizes. Unanswered questions include: - Where is the button placed? - What does the button look like (i.e. what is the icon?)? - What are the keyboard interaction/accessibility needs for this button - Providing a column header menu action for removing the space (same functionality as button mentioned above). - Providing a new table menu that would include an action item to remove the space. Questions include: - Where would this menu go visually? - Provide a new column sizing behavior that result in columns "collapsing" below their minimum width (column data would not be visible), ultimately resulting in no horizontal scrollable area. This behavior has a lot of unresolved corner cases, and ultimately would be quite expensive to implement. Recommendation is to not consider this as an option. <!--- The implementing team may build a list of tasks/sub-issues here: ## 📋 Tasks - [ ] This is a subtask of the feature. (It can be converted to an issue.) -->
1.0
Allow user to remove horizontal scroll space introduced by sizing table columns - ## 😯 Problem to Solve The current behavior for interactive column resizing allows for the creation of a horizontal scrollable area. At the moment, once created, the scrollable area can not be removed sans refreshing the page, which isn't great. It would be nice for us to provide the user with a means of removing the extra horizontal scrollable space. ## 💁 Proposed Solution There has already been some visual/interactive design for this problem. Some considered solutions include: - Providing a button that appears once there is a horizontal scrollbar present that, when pressed, removes any excess horizontal space, while maintaining existing column proportional sizes. Unanswered questions include: - Where is the button placed? - What does the button look like (i.e. what is the icon?)? - What are the keyboard interaction/accessibility needs for this button - Providing a column header menu action for removing the space (same functionality as button mentioned above). - Providing a new table menu that would include an action item to remove the space. Questions include: - Where would this menu go visually? - Provide a new column sizing behavior that result in columns "collapsing" below their minimum width (column data would not be visible), ultimately resulting in no horizontal scrollable area. This behavior has a lot of unresolved corner cases, and ultimately would be quite expensive to implement. Recommendation is to not consider this as an option. <!--- The implementing team may build a list of tasks/sub-issues here: ## 📋 Tasks - [ ] This is a subtask of the feature. (It can be converted to an issue.) -->
non_process
allow user to remove horizontal scroll space introduced by sizing table columns 😯 problem to solve the current behavior for interactive column resizing allows for the creation of a horizontal scrollable area at the moment once created the scrollable area can not be removed sans refreshing the page which isn t great it would be nice for us to provide the user with a means of removing the extra horizontal scrollable space 💁 proposed solution there has already been some visual interactive design for this problem some considered solutions include providing a button that appears once there is a horizontal scrollbar present that when pressed removes any excess horizontal space while maintaining existing column proportional sizes unanswered questions include where is the button placed what does the button look like i e what is the icon what are the keyboard interaction accessibility needs for this button providing a column header menu action for removing the space same functionality as button mentioned above providing a new table menu that would include an action item to remove the space questions include where would this menu go visually provide a new column sizing behavior that result in columns collapsing below their minimum width column data would not be visible ultimately resulting in no horizontal scrollable area this behavior has a lot of unresolved corner cases and ultimately would be quite expensive to implement recommendation is to not consider this as an option the implementing team may build a list of tasks sub issues here 📋 tasks this is a subtask of the feature it can be converted to an issue
0
526
2,998,345,894
IssuesEvent
2015-07-23 13:40:19
scieloorg/search-journals
https://api.github.com/repos/scieloorg/search-journals
closed
Busca não condiz com ambiente em produção
Processamento
Fiz uma busca por "AIDS" no resumo e não encontrou nenhum resultado. Fiz a mesma busca no portal atual e encontrou mais de 3000 resultados: http://search.scielo.org/?output=site&lang=pt&from=0&sort=&format=abstract&count=20&fb=&page=1&q=aids&index=ab&where=ORG&search_form_submit=Pesquisar
1.0
Busca não condiz com ambiente em produção - Fiz uma busca por "AIDS" no resumo e não encontrou nenhum resultado. Fiz a mesma busca no portal atual e encontrou mais de 3000 resultados: http://search.scielo.org/?output=site&lang=pt&from=0&sort=&format=abstract&count=20&fb=&page=1&q=aids&index=ab&where=ORG&search_form_submit=Pesquisar
process
busca não condiz com ambiente em produção fiz uma busca por aids no resumo e não encontrou nenhum resultado fiz a mesma busca no portal atual e encontrou mais de resultados
1
18,092
24,118,444,875
IssuesEvent
2022-09-20 16:29:21
keras-team/keras-cv
https://api.github.com/repos/keras-team/keras-cv
closed
Add augment_bounding_boxes support to AugMix layer
contribution-welcome preprocessing
The augment_bounding_boxes should be implemented for AugMix Layer in keras_cv. The PR should contain implementation, test scripts and a demo script to verify implementation. Example code for implementing augment_bounding_boxes() can be found here - https://github.com/keras-team/keras-cv/blob/master/keras_cv/layers/preprocessing/random_flip.py#:~:text=def%20augment_bounding_boxes(,)%3A - https://github.com/keras-team/keras-cv/blob/master/keras_cv/layers/preprocessing/random_rotation.py#:~:text=def%20augment_image(self%2C%20image%2C%20transformation%2C%20**kwargs)%3A - The implementations can be verified using demo utils in keras_cv.bounding_box - Example of demo script can be found here : https://github.com/keras-team/keras-cv/blob/master/examples/layers/preprocessing/bounding_box/random_rotation_demo.py
1.0
Add augment_bounding_boxes support to AugMix layer - The augment_bounding_boxes should be implemented for AugMix Layer in keras_cv. The PR should contain implementation, test scripts and a demo script to verify implementation. Example code for implementing augment_bounding_boxes() can be found here - https://github.com/keras-team/keras-cv/blob/master/keras_cv/layers/preprocessing/random_flip.py#:~:text=def%20augment_bounding_boxes(,)%3A - https://github.com/keras-team/keras-cv/blob/master/keras_cv/layers/preprocessing/random_rotation.py#:~:text=def%20augment_image(self%2C%20image%2C%20transformation%2C%20**kwargs)%3A - The implementations can be verified using demo utils in keras_cv.bounding_box - Example of demo script can be found here : https://github.com/keras-team/keras-cv/blob/master/examples/layers/preprocessing/bounding_box/random_rotation_demo.py
process
add augment bounding boxes support to augmix layer the augment bounding boxes should be implemented for augmix layer in keras cv the pr should contain implementation test scripts and a demo script to verify implementation example code for implementing augment bounding boxes can be found here the implementations can be verified using demo utils in keras cv bounding box example of demo script can be found here
1
51,013
13,612,770,954
IssuesEvent
2020-09-23 10:49:12
jgeraigery/angular-breadcrumb
https://api.github.com/repos/jgeraigery/angular-breadcrumb
opened
CVE-2020-7608 (Medium) detected in multiple libraries
security vulnerability
## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-9.0.2.tgz</b>, <b>yargs-parser-5.0.0.tgz</b>, <b>yargs-parser-10.1.0.tgz</b>, <b>yargs-parser-7.0.0.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-9.0.2.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/copyfiles/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - copyfiles-2.0.0.tgz (Root Library) - yargs-11.0.0.tgz - :x: **yargs-parser-9.0.2.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/sass-graph/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.12.4.tgz (Root Library) - node-sass-4.10.0.tgz - sass-graph-2.2.4.tgz - yargs-7.1.0.tgz - :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.12.4.tgz (Root Library) - webpack-dev-server-3.1.14.tgz - yargs-12.0.2.tgz - :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - compiler-cli-7.2.3.tgz (Root Library) - yargs-9.0.1.tgz - :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/angular-breadcrumb/commit/7d59b99de33040c4ee0da1186de492a71c07d62d">7d59b99de33040c4ee0da1186de492a71c07d62d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"9.0.2","isTransitiveDependency":true,"dependencyTree":"copyfiles:2.0.0;yargs:11.0.0;yargs-parser:9.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"5.0.0","isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.12.4;node-sass:4.10.0;sass-graph:2.2.4;yargs:7.1.0;yargs-parser:5.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"10.1.0","isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.12.4;webpack-dev-server:3.1.14;yargs:12.0.2;yargs-parser:10.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"7.0.0","isTransitiveDependency":true,"dependencyTree":"@angular/compiler-cli:7.2.3;yargs:9.0.1;yargs-parser:7.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7608 (Medium) detected in multiple libraries - ## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-9.0.2.tgz</b>, <b>yargs-parser-5.0.0.tgz</b>, <b>yargs-parser-10.1.0.tgz</b>, <b>yargs-parser-7.0.0.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-9.0.2.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/copyfiles/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - copyfiles-2.0.0.tgz (Root Library) - yargs-11.0.0.tgz - :x: **yargs-parser-9.0.2.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/sass-graph/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.12.4.tgz (Root Library) - node-sass-4.10.0.tgz - sass-graph-2.2.4.tgz - yargs-7.1.0.tgz - :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.12.4.tgz (Root Library) - webpack-dev-server-3.1.14.tgz - yargs-12.0.2.tgz - :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p> <p>Path to dependency file: angular-breadcrumb/package.json</p> <p>Path to vulnerable library: angular-breadcrumb/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - compiler-cli-7.2.3.tgz (Root Library) - yargs-9.0.1.tgz - :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/angular-breadcrumb/commit/7d59b99de33040c4ee0da1186de492a71c07d62d">7d59b99de33040c4ee0da1186de492a71c07d62d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"9.0.2","isTransitiveDependency":true,"dependencyTree":"copyfiles:2.0.0;yargs:11.0.0;yargs-parser:9.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"5.0.0","isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.12.4;node-sass:4.10.0;sass-graph:2.2.4;yargs:7.1.0;yargs-parser:5.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"10.1.0","isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.12.4;webpack-dev-server:3.1.14;yargs:12.0.2;yargs-parser:10.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"7.0.0","isTransitiveDependency":true,"dependencyTree":"@angular/compiler-cli:7.2.3;yargs:9.0.1;yargs-parser:7.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file angular breadcrumb package json path to vulnerable library angular breadcrumb node modules copyfiles node modules yargs parser package json dependency hierarchy copyfiles tgz root library yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file angular breadcrumb package json path to vulnerable library angular breadcrumb node modules sass graph node modules yargs parser package json dependency hierarchy build angular tgz root library node sass tgz sass graph tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file angular breadcrumb package json path to vulnerable library angular breadcrumb node modules webpack dev server node modules yargs parser package json dependency hierarchy build angular tgz root library webpack dev server tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file angular breadcrumb package json path to vulnerable library angular breadcrumb node modules yargs parser package json dependency hierarchy compiler cli tgz root library yargs tgz x yargs parser tgz vulnerable library found in head commit a href found in base branch master vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload vulnerabilityurl
0
49,582
6,033,529,701
IssuesEvent
2017-06-09 08:35:14
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
[FVT]update script pre_deploy_sn to support PostgreSQL database in automation test bucket
component:test priority:low sprint2 status:pending test:testcase_added
We need to support both Mysql and PostgreSQL in automaton bucket. Need to update pre_deploy_sn to run PostgreSQL setup script. ``` $cmd = "XCATMYSQLADMIN_PW=12345 XCATMYSQLROOT_PW=12345 /opt/xcat/bin/mysqlsetup -i -V"; runcmd("$cmd"); $cmd = "echo \"GRANT ALL on xcatdb.* TO xcatadmin@\'%\' IDENTIFIED BY \'12345\'\;\" | mysql -u root -p12345"; runcmd("$cmd"); ```
3.0
[FVT]update script pre_deploy_sn to support PostgreSQL database in automation test bucket - We need to support both Mysql and PostgreSQL in automaton bucket. Need to update pre_deploy_sn to run PostgreSQL setup script. ``` $cmd = "XCATMYSQLADMIN_PW=12345 XCATMYSQLROOT_PW=12345 /opt/xcat/bin/mysqlsetup -i -V"; runcmd("$cmd"); $cmd = "echo \"GRANT ALL on xcatdb.* TO xcatadmin@\'%\' IDENTIFIED BY \'12345\'\;\" | mysql -u root -p12345"; runcmd("$cmd"); ```
non_process
update script pre deploy sn to support postgresql database in automation test bucket we need to support both mysql and postgresql in automaton bucket need to update pre deploy sn to run postgresql setup script cmd xcatmysqladmin pw xcatmysqlroot pw opt xcat bin mysqlsetup i v runcmd cmd cmd echo grant all on xcatdb to xcatadmin identified by mysql u root runcmd cmd
0
723,223
24,889,920,229
IssuesEvent
2022-10-28 11:02:50
robotframework/robotframework
https://api.github.com/repos/robotframework/robotframework
closed
Libdoc's `DocumentationBuilder` doesn't anymore work with resource files with `.robot` extension
bug priority: medium
When support for generating documentation for suite files was added (#4493), the `DocumentationBuilder` factory method was changed so that it returns a `SuiteDocBuilder` with all files having a `.robot` extension. This builder only works with suite files and not with resource files. The `LibraryDocumentation` factory method that uses `DocumentationBuilder` handles this problem internally similarly as it handles `DocumentationBuilder` not supporting resource files in PYTHONPATH so this doesn't affect Libdoc itself. It does, however, affect direct usage of `DocumentationBuilder`. Changes like this are always annoying, but in this case this isn't too severe because `DocumentationBuilder` isn't really part of the public API and `LibraryDocumentation` should be used instead of it. Alternatively, it's possible to use `ResourceDocBuilder` explicitly when working with resource files. The initial plan was to just document this behavior (#4517), but it is actually easy to fix this by making `DocumentationBuilder` a class instead of a factory method. A related problem is that Libdoc's public API isn't properly defined. A separate issue needs to be submitted about that. We can then explicitly recommend `LibraryDocumentation`.
1.0
Libdoc's `DocumentationBuilder` doesn't anymore work with resource files with `.robot` extension - When support for generating documentation for suite files was added (#4493), the `DocumentationBuilder` factory method was changed so that it returns a `SuiteDocBuilder` with all files having a `.robot` extension. This builder only works with suite files and not with resource files. The `LibraryDocumentation` factory method that uses `DocumentationBuilder` handles this problem internally similarly as it handles `DocumentationBuilder` not supporting resource files in PYTHONPATH so this doesn't affect Libdoc itself. It does, however, affect direct usage of `DocumentationBuilder`. Changes like this are always annoying, but in this case this isn't too severe because `DocumentationBuilder` isn't really part of the public API and `LibraryDocumentation` should be used instead of it. Alternatively, it's possible to use `ResourceDocBuilder` explicitly when working with resource files. The initial plan was to just document this behavior (#4517), but it is actually easy to fix this by making `DocumentationBuilder` a class instead of a factory method. A related problem is that Libdoc's public API isn't properly defined. A separate issue needs to be submitted about that. We can then explicitly recommend `LibraryDocumentation`.
non_process
libdoc s documentationbuilder doesn t anymore work with resource files with robot extension when support for generating documentation for suite files was added the documentationbuilder factory method was changed so that it returns a suitedocbuilder with all files having a robot extension this builder only works with suite files and not with resource files the librarydocumentation factory method that uses documentationbuilder handles this problem internally similarly as it handles documentationbuilder not supporting resource files in pythonpath so this doesn t affect libdoc itself it does however affect direct usage of documentationbuilder changes like this are always annoying but in this case this isn t too severe because documentationbuilder isn t really part of the public api and librarydocumentation should be used instead of it alternatively it s possible to use resourcedocbuilder explicitly when working with resource files the initial plan was to just document this behavior but it is actually easy to fix this by making documentationbuilder a class instead of a factory method a related problem is that libdoc s public api isn t properly defined a separate issue needs to be submitted about that we can then explicitly recommend librarydocumentation
0
18,562
24,555,700,393
IssuesEvent
2022-10-12 15:42:16
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] Enrollment flow > Data sharing permission > Participant is able to select both data sharing options present on the screen
Bug P0 iOS Process: Fixed Process: Tested dev
AR: Enrollment flow > Data sharing permission > Participant is able to select both data sharing options present on the screen ER: Participants should be able to select only one option at a time on the data sharing permission screen ![image](https://user-images.githubusercontent.com/71445210/188405882-79fec2d7-5769-4ecd-a78b-73ea2d98849c.png)
2.0
[iOS] Enrollment flow > Data sharing permission > Participant is able to select both data sharing options present on the screen - AR: Enrollment flow > Data sharing permission > Participant is able to select both data sharing options present on the screen ER: Participants should be able to select only one option at a time on the data sharing permission screen ![image](https://user-images.githubusercontent.com/71445210/188405882-79fec2d7-5769-4ecd-a78b-73ea2d98849c.png)
process
enrollment flow data sharing permission participant is able to select both data sharing options present on the screen ar enrollment flow data sharing permission participant is able to select both data sharing options present on the screen er participants should be able to select only one option at a time on the data sharing permission screen
1
79,343
3,534,867,832
IssuesEvent
2016-01-16 02:49:39
Valhalla-Gaming/Tracker
https://api.github.com/repos/Valhalla-Gaming/Tracker
closed
Warlock Ember gains
Class-Warlock Priority-Critical Type-Spell
so just tested destor warlocks after the recent update, and i noticed that if i have a dot heal on me each tick grants 1 Ember and i can literally spam chaol bolts 24/7 if i have a HoT on me, tested again seems all heals on me grants Embers
1.0
Warlock Ember gains - so just tested destor warlocks after the recent update, and i noticed that if i have a dot heal on me each tick grants 1 Ember and i can literally spam chaol bolts 24/7 if i have a HoT on me, tested again seems all heals on me grants Embers
non_process
warlock ember gains so just tested destor warlocks after the recent update and i noticed that if i have a dot heal on me each tick grants ember and i can literally spam chaol bolts if i have a hot on me tested again seems all heals on me grants embers
0
6,123
8,996,401,689
IssuesEvent
2019-02-02 01:10:52
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Expand Process.Kill to Optionally Kill a Process Tree
api-approved area-System.Diagnostics.Process
## Issue .Net Standard does not provide a means to kill a process tree (that is, a given process and all of its child/descendant processes). In the Windows world, this can be done by assembling the list of descendant processes to terminate using [WMI's Win32_Process management object](https://stackoverflow.com/a/7189381/117424), [performance counters](https://stackoverflow.com/a/2336322/117424) or [P/Invoke](https://stackoverflow.com/a/3346055/117424), then killing each one, then killing the parent/root process. However, since .Net Standard doesn't provide a way to enumerate child/descendant processes, it's impossible for a user to directly implement this approach in a cross-platform way using only .Net Standard. If the ability to enumerate child processes were added (as #25855 proposes), there'd still be complexities involved with using the child process list to kill a process tree as care is needed to ensure that unrelated processes aren't inadvertently terminated because of process ID reuse. This proposal saves the developer from needing to worry about such complexities and instead provides a simple, one-line way to terminate a process tree. ## Rationale Sometimes, it's necessary to end a process tree. :-) For example, [`Microsoft.DotNet.Tools.Test.Utilities.ProcessExtension`](https://github.com/dotnet/cli/blob/35e661bbb3b2f0943590cb63747ae53a68e6eb5d/test/Microsoft.DotNet.Tools.Tests.Utilities/Extensions/ProcessExtensions.cs#L27) relies on different code paths to handle killing a process tree on Windows and Linux. Both code paths involve invoking one or more command-line processes. If this proposal were implemented, presumably this logic could be replaced with a single-line kill all method call followed possibly by a call to [`WaithForExit(int milliseconds)`](https://docs.microsoft.com/en-us/dotnet/api/system.diagnostics.process.waitforexit?view=netstandard-2.0#System_Diagnostics_Process_WaitForExit_System_Int32_). No multiple code paths. No need to invoke command-line processes. Another example: A project I'm currently working on uses integration tests that use [`Process`](https://docs.microsoft.com/en-us/dotnet/api/system.diagnostics.process?view=netstandard-2.0) to call `dotnet run` to start a server process. When the test has completed, it's possible to use Process's `Kill()` to end the parent dotnet process. However, this leaves child processes still running—which causes problems. ## Proposal Extend `Process` as follows: **Original (ignore—left for reference purposes):** ```c# public class Process : .... { public void Kill(bool entireProcessTree = false); // false = existing behavior of Kill() } ``` **Revision (ignore—left for reference purposes):** ```c# public enum ProcessGroup { CurrentProcess, EntireTree } public class Process : .... { public void Kill(ProcessGroup group = ProcessGroup.CurrentProcess) { } // default = existing behavior of Kill() } ``` **Approved Version:** ``` public class Process : .... { public void Kill(bool entireProcessTree); } ``` With this change, `Kill` will behave exactly as it does at current if it is called with no arguments or with `false` as its argument. However, when called with `true` as its argument, in a platform-appropriate way all descendant processes will be killed, followed by the process represented by the current `Process` instance. [edited by @danmosemsft to add c# formatting only]
1.0
Expand Process.Kill to Optionally Kill a Process Tree - ## Issue .Net Standard does not provide a means to kill a process tree (that is, a given process and all of its child/descendant processes). In the Windows world, this can be done by assembling the list of descendant processes to terminate using [WMI's Win32_Process management object](https://stackoverflow.com/a/7189381/117424), [performance counters](https://stackoverflow.com/a/2336322/117424) or [P/Invoke](https://stackoverflow.com/a/3346055/117424), then killing each one, then killing the parent/root process. However, since .Net Standard doesn't provide a way to enumerate child/descendant processes, it's impossible for a user to directly implement this approach in a cross-platform way using only .Net Standard. If the ability to enumerate child processes were added (as #25855 proposes), there'd still be complexities involved with using the child process list to kill a process tree as care is needed to ensure that unrelated processes aren't inadvertently terminated because of process ID reuse. This proposal saves the developer from needing to worry about such complexities and instead provides a simple, one-line way to terminate a process tree. ## Rationale Sometimes, it's necessary to end a process tree. :-) For example, [`Microsoft.DotNet.Tools.Test.Utilities.ProcessExtension`](https://github.com/dotnet/cli/blob/35e661bbb3b2f0943590cb63747ae53a68e6eb5d/test/Microsoft.DotNet.Tools.Tests.Utilities/Extensions/ProcessExtensions.cs#L27) relies on different code paths to handle killing a process tree on Windows and Linux. Both code paths involve invoking one or more command-line processes. If this proposal were implemented, presumably this logic could be replaced with a single-line kill all method call followed possibly by a call to [`WaithForExit(int milliseconds)`](https://docs.microsoft.com/en-us/dotnet/api/system.diagnostics.process.waitforexit?view=netstandard-2.0#System_Diagnostics_Process_WaitForExit_System_Int32_). No multiple code paths. No need to invoke command-line processes. Another example: A project I'm currently working on uses integration tests that use [`Process`](https://docs.microsoft.com/en-us/dotnet/api/system.diagnostics.process?view=netstandard-2.0) to call `dotnet run` to start a server process. When the test has completed, it's possible to use Process's `Kill()` to end the parent dotnet process. However, this leaves child processes still running—which causes problems. ## Proposal Extend `Process` as follows: **Original (ignore—left for reference purposes):** ```c# public class Process : .... { public void Kill(bool entireProcessTree = false); // false = existing behavior of Kill() } ``` **Revision (ignore—left for reference purposes):** ```c# public enum ProcessGroup { CurrentProcess, EntireTree } public class Process : .... { public void Kill(ProcessGroup group = ProcessGroup.CurrentProcess) { } // default = existing behavior of Kill() } ``` **Approved Version:** ``` public class Process : .... { public void Kill(bool entireProcessTree); } ``` With this change, `Kill` will behave exactly as it does at current if it is called with no arguments or with `false` as its argument. However, when called with `true` as its argument, in a platform-appropriate way all descendant processes will be killed, followed by the process represented by the current `Process` instance. [edited by @danmosemsft to add c# formatting only]
process
expand process kill to optionally kill a process tree issue net standard does not provide a means to kill a process tree that is a given process and all of its child descendant processes in the windows world this can be done by assembling the list of descendant processes to terminate using or then killing each one then killing the parent root process however since net standard doesn t provide a way to enumerate child descendant processes it s impossible for a user to directly implement this approach in a cross platform way using only net standard if the ability to enumerate child processes were added as proposes there d still be complexities involved with using the child process list to kill a process tree as care is needed to ensure that unrelated processes aren t inadvertently terminated because of process id reuse this proposal saves the developer from needing to worry about such complexities and instead provides a simple one line way to terminate a process tree rationale sometimes it s necessary to end a process tree for example relies on different code paths to handle killing a process tree on windows and linux both code paths involve invoking one or more command line processes if this proposal were implemented presumably this logic could be replaced with a single line kill all method call followed possibly by a call to no multiple code paths no need to invoke command line processes another example a project i m currently working on uses integration tests that use to call dotnet run to start a server process when the test has completed it s possible to use process s kill to end the parent dotnet process however this leaves child processes still running—which causes problems proposal extend process as follows original ignore—left for reference purposes c public class process public void kill bool entireprocesstree false false existing behavior of kill revision ignore—left for reference purposes c public enum processgroup currentprocess entiretree public class process public void kill processgroup group processgroup currentprocess default existing behavior of kill approved version public class process public void kill bool entireprocesstree with this change kill will behave exactly as it does at current if it is called with no arguments or with false as its argument however when called with true as its argument in a platform appropriate way all descendant processes will be killed followed by the process represented by the current process instance
1
776,884
27,264,742,308
IssuesEvent
2023-02-22 17:10:27
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
opened
SA-17(4): Developer Security Architecture And Design | Informal Correspondence
Priority: P3 ITSG-33 Suggested Assignment: IT Projects Class: Management Control: SA-17
# Control Definition DEVELOPER SECURITY ARCHITECTURE AND DESIGN | INFORMAL CORRESPONDENCE The organization requires the developer of the information system, system component, or information system service to: (a) Produce, as an integral part of the development process, an informal descriptive top-level specification that specifies the interfaces to security-relevant hardware, software, and firmware in terms of exceptions, error messages, and effects; (b) Show via [Selection: informal demonstration, convincing argument with formal methods as feasible] that the descriptive top-level specification is consistent with the formal policy model; (c) Show via informal demonstration, that the descriptive top-level specification completely covers the interfaces to security-relevant hardware, software, and firmware; (d) Show that the descriptive top-level specification is an accurate description of the interfaces to security-relevant hardware, software, and firmware; and (e) Describe the security-relevant hardware, software, and firmware mechanisms not addressed in the descriptive top-level specification but strictly internal to the security-relevant hardware, software, and firmware. # Class Management # Supplemental Guidance Correspondence is an important part of the assurance gained through modeling. It demonstrates that the implementation is an accurate transformation of the model, and that any additional code or implementation details present has no impact on the behaviours or policies being modeled. Consistency between the descriptive top-level specification (i.e., high-level/low-level design) and the formal policy model is generally not amenable to being fully proven. Therefore, a combination of formal/informal methods may be needed to show such consistency. Hardware, software, and firmware mechanisms strictly internal to security-relevant hardware, software, and firmware include, for example, mapping registers and direct memory input/output. Related control: SA-5. # General Guide Apply to custom developed systems or components. # Suggested Assignment IT Projects # Support Teams IT Security Function
1.0
SA-17(4): Developer Security Architecture And Design | Informal Correspondence - # Control Definition DEVELOPER SECURITY ARCHITECTURE AND DESIGN | INFORMAL CORRESPONDENCE The organization requires the developer of the information system, system component, or information system service to: (a) Produce, as an integral part of the development process, an informal descriptive top-level specification that specifies the interfaces to security-relevant hardware, software, and firmware in terms of exceptions, error messages, and effects; (b) Show via [Selection: informal demonstration, convincing argument with formal methods as feasible] that the descriptive top-level specification is consistent with the formal policy model; (c) Show via informal demonstration, that the descriptive top-level specification completely covers the interfaces to security-relevant hardware, software, and firmware; (d) Show that the descriptive top-level specification is an accurate description of the interfaces to security-relevant hardware, software, and firmware; and (e) Describe the security-relevant hardware, software, and firmware mechanisms not addressed in the descriptive top-level specification but strictly internal to the security-relevant hardware, software, and firmware. # Class Management # Supplemental Guidance Correspondence is an important part of the assurance gained through modeling. It demonstrates that the implementation is an accurate transformation of the model, and that any additional code or implementation details present has no impact on the behaviours or policies being modeled. Consistency between the descriptive top-level specification (i.e., high-level/low-level design) and the formal policy model is generally not amenable to being fully proven. Therefore, a combination of formal/informal methods may be needed to show such consistency. Hardware, software, and firmware mechanisms strictly internal to security-relevant hardware, software, and firmware include, for example, mapping registers and direct memory input/output. Related control: SA-5. # General Guide Apply to custom developed systems or components. # Suggested Assignment IT Projects # Support Teams IT Security Function
non_process
sa developer security architecture and design informal correspondence control definition developer security architecture and design informal correspondence the organization requires the developer of the information system system component or information system service to a produce as an integral part of the development process an informal descriptive top level specification that specifies the interfaces to security relevant hardware software and firmware in terms of exceptions error messages and effects b show via that the descriptive top level specification is consistent with the formal policy model c show via informal demonstration that the descriptive top level specification completely covers the interfaces to security relevant hardware software and firmware d show that the descriptive top level specification is an accurate description of the interfaces to security relevant hardware software and firmware and e describe the security relevant hardware software and firmware mechanisms not addressed in the descriptive top level specification but strictly internal to the security relevant hardware software and firmware class management supplemental guidance correspondence is an important part of the assurance gained through modeling it demonstrates that the implementation is an accurate transformation of the model and that any additional code or implementation details present has no impact on the behaviours or policies being modeled consistency between the descriptive top level specification i e high level low level design and the formal policy model is generally not amenable to being fully proven therefore a combination of formal informal methods may be needed to show such consistency hardware software and firmware mechanisms strictly internal to security relevant hardware software and firmware include for example mapping registers and direct memory input output related control sa general guide apply to custom developed systems or components suggested assignment it projects support teams it security function
0
6,923
10,082,811,085
IssuesEvent
2019-07-25 12:14:54
linnovate/root
https://api.github.com/repos/linnovate/root
closed
Multipale selecte does not work on the tasks I following to
Fixed Process bug bug critical
- I have marked one task, in fact all have been marked
1.0
Multipale selecte does not work on the tasks I following to - - I have marked one task, in fact all have been marked
process
multipale selecte does not work on the tasks i following to i have marked one task in fact all have been marked
1
41,296
8,957,841,021
IssuesEvent
2019-01-27 08:41:34
opencodeiiita/Competitive_Coding
https://api.github.com/repos/opencodeiiita/Competitive_Coding
closed
Logical error
Everyone OpenCode19 Skilled: 20 points bug open
Task 3, Issue 3: Find and fix the logical error by modifying not more than one line.
1.0
Logical error - Task 3, Issue 3: Find and fix the logical error by modifying not more than one line.
non_process
logical error task issue find and fix the logical error by modifying not more than one line
0
261,488
27,809,782,609
IssuesEvent
2023-03-18 01:43:07
madhans23/linux-4.1.15
https://api.github.com/repos/madhans23/linux-4.1.15
closed
CVE-2022-21504 (Medium) detected in linux-stable-rtv4.1.33, linux-yocto-4.1v4.1.17 - autoclosed
Mend: dependency security vulnerability
## CVE-2022-21504 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-yocto-4.1v4.1.17</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The code in UEK6 U3 was missing an appropiate file descriptor count to be missing. This resulted in a use count error that allowed a file descriptor to a socket to be closed and freed while it was still in use by another portion of the kernel. An attack with local access can operate on the socket, and cause a denial of service. CVSS 3.1 Base Score 5.5 (Availability impacts). CVSS Vector: (CVSS:3.1/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H). <p>Publish Date: 2022-06-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21504>CVE-2022-21504</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-14</p> <p>Fix Resolution: v5.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-21504 (Medium) detected in linux-stable-rtv4.1.33, linux-yocto-4.1v4.1.17 - autoclosed - ## CVE-2022-21504 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-yocto-4.1v4.1.17</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The code in UEK6 U3 was missing an appropiate file descriptor count to be missing. This resulted in a use count error that allowed a file descriptor to a socket to be closed and freed while it was still in use by another portion of the kernel. An attack with local access can operate on the socket, and cause a denial of service. CVSS 3.1 Base Score 5.5 (Availability impacts). CVSS Vector: (CVSS:3.1/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H). <p>Publish Date: 2022-06-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21504>CVE-2022-21504</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-14</p> <p>Fix Resolution: v5.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable linux yocto autoclosed cve medium severity vulnerability vulnerable libraries linux stable linux yocto vulnerability details the code in was missing an appropiate file descriptor count to be missing this resulted in a use count error that allowed a file descriptor to a socket to be closed and freed while it was still in use by another portion of the kernel an attack with local access can operate on the socket and cause a denial of service cvss base score availability impacts cvss vector cvss av l ac l pr l ui n s u c n i n a h publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend
0
2,895
2,607,964,946
IssuesEvent
2015-02-26 00:41:55
chrsmithdemos/leveldb
https://api.github.com/repos/chrsmithdemos/leveldb
closed
Provide "close" interface to leveldb
auto-migrated Priority-Medium Type-Defect
``` According to the API document, for Closing A Database, just delete the database object. Example: ... open the db as described above ... ... do something with db ... delete db; It might leads to some encapsulation error. For example, if there were a storage encapsuator with "clear" interface as well as iterator interface, and within the implementation of "clear": void clear() { delete pLeveldb_; } If "clear" interface is called before iterator has been destroyed, it will be easy to lead to corruption since the leveldb instance contained by the iterator has been obsolete. ``` ----- Original issue reported on code.google.com by `yingfeng.zhang` on 3 Feb 2012 at 1:25
1.0
Provide "close" interface to leveldb - ``` According to the API document, for Closing A Database, just delete the database object. Example: ... open the db as described above ... ... do something with db ... delete db; It might leads to some encapsulation error. For example, if there were a storage encapsuator with "clear" interface as well as iterator interface, and within the implementation of "clear": void clear() { delete pLeveldb_; } If "clear" interface is called before iterator has been destroyed, it will be easy to lead to corruption since the leveldb instance contained by the iterator has been obsolete. ``` ----- Original issue reported on code.google.com by `yingfeng.zhang` on 3 Feb 2012 at 1:25
non_process
provide close interface to leveldb according to the api document for closing a database just delete the database object example open the db as described above do something with db delete db it might leads to some encapsulation error for example if there were a storage encapsuator with clear interface as well as iterator interface and within the implementation of clear void clear delete pleveldb if clear interface is called before iterator has been destroyed it will be easy to lead to corruption since the leveldb instance contained by the iterator has been obsolete original issue reported on code google com by yingfeng zhang on feb at
0
257,472
8,137,809,664
IssuesEvent
2018-08-20 13:03:26
themuseblockchain/Muse-Source
https://api.github.com/repos/themuseblockchain/Muse-Source
closed
Provide a way to pay to content
enhancement hardfork high priority
Currently the smart contract made up by a piece of content is only used for paying out content rewards. It would be nice if there was a way to pay an amount of money to a piece of content that is automatically split among the content's payees.
1.0
Provide a way to pay to content - Currently the smart contract made up by a piece of content is only used for paying out content rewards. It would be nice if there was a way to pay an amount of money to a piece of content that is automatically split among the content's payees.
non_process
provide a way to pay to content currently the smart contract made up by a piece of content is only used for paying out content rewards it would be nice if there was a way to pay an amount of money to a piece of content that is automatically split among the content s payees
0
456,494
13,150,825,940
IssuesEvent
2020-08-09 13:43:20
chrisjsewell/docutils
https://api.github.com/repos/chrisjsewell/docutils
closed
UTF8 in stylesheets [SF:bugs:79]
bugs closed-fixed priority-5
author: sonderblade created: 2007-04-02 10:44:35 assigned: None SF_url: https://sourceforge.net/p/docutils/bugs/79 rst2html.py does not like this ":Author: Björn Lindqvist" in stylesheets: Traceback \(most recent call last\): File "C:\Python25\rst2html.py", line 25, in <module> publish\_cmdline\(writer\_name='html', description=description\) File "C:\Python25\lib\site-packages\docutils\core.py", line 335, in publish\_cmdline config\_section=config\_section, enable\_exit\_status=enable\_exit\_status\) File "C:\Python25\lib\site-packages\docutils\core.py", line 217, in publish self.report\_Exception\(error\) File "C:\Python25\lib\site-packages\docutils\core.py", line 256, in report\_Exception self.report\_UnicodeError\(error\) File "C:\Python25\lib\site-packages\docutils\core.py", line 289, in report\_UnicodeError % \(data.encode\('ascii', 'xmlcharrefreplace'\), UnicodeDecodeError: 'ascii' codec can't decode byte 0xf6 in position 0: ordinal not in range\(128\) Trying to fiddle with input encoding settings does not help. --- commenter: milde posted: 2009-09-07 08:58:13 title: #79 UTF8 in stylesheets Fixed in both html4css1 and latex2e writer. --- commenter: milde posted: 2009-09-07 08:58:15 title: #79 UTF8 in stylesheets - **status**: open --> open-fixed --- commenter: milde posted: 2009-09-07 09:08:39 title: #79 UTF8 in stylesheets Fixed in SVN --- commenter: milde posted: 2009-09-10 11:41:27 title: #79 UTF8 in stylesheets - **status**: open-fixed --> closed-fixed
1.0
UTF8 in stylesheets [SF:bugs:79] - author: sonderblade created: 2007-04-02 10:44:35 assigned: None SF_url: https://sourceforge.net/p/docutils/bugs/79 rst2html.py does not like this ":Author: Björn Lindqvist" in stylesheets: Traceback \(most recent call last\): File "C:\Python25\rst2html.py", line 25, in <module> publish\_cmdline\(writer\_name='html', description=description\) File "C:\Python25\lib\site-packages\docutils\core.py", line 335, in publish\_cmdline config\_section=config\_section, enable\_exit\_status=enable\_exit\_status\) File "C:\Python25\lib\site-packages\docutils\core.py", line 217, in publish self.report\_Exception\(error\) File "C:\Python25\lib\site-packages\docutils\core.py", line 256, in report\_Exception self.report\_UnicodeError\(error\) File "C:\Python25\lib\site-packages\docutils\core.py", line 289, in report\_UnicodeError % \(data.encode\('ascii', 'xmlcharrefreplace'\), UnicodeDecodeError: 'ascii' codec can't decode byte 0xf6 in position 0: ordinal not in range\(128\) Trying to fiddle with input encoding settings does not help. --- commenter: milde posted: 2009-09-07 08:58:13 title: #79 UTF8 in stylesheets Fixed in both html4css1 and latex2e writer. --- commenter: milde posted: 2009-09-07 08:58:15 title: #79 UTF8 in stylesheets - **status**: open --> open-fixed --- commenter: milde posted: 2009-09-07 09:08:39 title: #79 UTF8 in stylesheets Fixed in SVN --- commenter: milde posted: 2009-09-10 11:41:27 title: #79 UTF8 in stylesheets - **status**: open-fixed --> closed-fixed
non_process
in stylesheets author sonderblade created assigned none sf url py does not like this author björn lindqvist in stylesheets traceback most recent call last file c py line in publish cmdline writer name html description description file c lib site packages docutils core py line in publish cmdline config section config section enable exit status enable exit status file c lib site packages docutils core py line in publish self report exception error file c lib site packages docutils core py line in report exception self report unicodeerror error file c lib site packages docutils core py line in report unicodeerror data encode ascii xmlcharrefreplace unicodedecodeerror ascii codec can t decode byte in position ordinal not in range trying to fiddle with input encoding settings does not help commenter milde posted title in stylesheets fixed in both and writer commenter milde posted title in stylesheets status open open fixed commenter milde posted title in stylesheets fixed in svn commenter milde posted title in stylesheets status open fixed closed fixed
0
123,466
26,262,570,445
IssuesEvent
2023-01-06 09:23:32
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[Mono] Remainder example fails for LLVMFullAot
area-Codegen-AOT-mono
```csharp // Licensed to the .NET Foundation under one or more agreements. // The .NET Foundation licenses this file to you under the MIT license. public class Program { public static ulong[,] s_1; public static int Main() { // This should not assert. try { ushort vr10 = default(ushort); bool vr11 = 0 < ((s_1[0, 0] * (uint)(0 / vr10)) % 1); } catch {} return 100; } } ```
1.0
[Mono] Remainder example fails for LLVMFullAot - ```csharp // Licensed to the .NET Foundation under one or more agreements. // The .NET Foundation licenses this file to you under the MIT license. public class Program { public static ulong[,] s_1; public static int Main() { // This should not assert. try { ushort vr10 = default(ushort); bool vr11 = 0 < ((s_1[0, 0] * (uint)(0 / vr10)) % 1); } catch {} return 100; } } ```
non_process
remainder example fails for llvmfullaot csharp licensed to the net foundation under one or more agreements the net foundation licenses this file to you under the mit license public class program public static ulong s public static int main this should not assert try ushort default ushort bool s uint catch return
0
20,165
26,719,010,759
IssuesEvent
2023-01-28 22:32:18
bisq-network/proposals
https://api.github.com/repos/bisq-network/proposals
closed
Set the minimum payout at mediation to be 5% of trade amount
was:approved a:proposal re:processes
## Proposal Currently the minimum payout at mediation is [2.5% of trade amount](https://github.com/bisq-network/bisq/pull/6362). This proposal suggests to increase this amount to be 5% of trade amount. ### Reasoning With the introduction [do not pay out the security deposit of the trade peer to the arbitration case winner](https://github.com/bisq-network/proposals/issues/386) there needs to be a bigger incentive for the penalized party to accept the mediation proposal to prevent the trade entering arbitration. This is because if the trade enters arbitration it will cause a delayed payout transaction and result in the penalized party likely not getting any reimbursement, and the non-penalized party likely not getting any compensation as a result. ### Outcome Make the penalized party more likely to accept mediation by incentivizing them to receive a minimum of 5% of the trade amount (currently twice the minimum they can receive). ### Solution Update the minimum payout a trader can receive to be 5% of the trade amount.
1.0
Set the minimum payout at mediation to be 5% of trade amount - ## Proposal Currently the minimum payout at mediation is [2.5% of trade amount](https://github.com/bisq-network/bisq/pull/6362). This proposal suggests to increase this amount to be 5% of trade amount. ### Reasoning With the introduction [do not pay out the security deposit of the trade peer to the arbitration case winner](https://github.com/bisq-network/proposals/issues/386) there needs to be a bigger incentive for the penalized party to accept the mediation proposal to prevent the trade entering arbitration. This is because if the trade enters arbitration it will cause a delayed payout transaction and result in the penalized party likely not getting any reimbursement, and the non-penalized party likely not getting any compensation as a result. ### Outcome Make the penalized party more likely to accept mediation by incentivizing them to receive a minimum of 5% of the trade amount (currently twice the minimum they can receive). ### Solution Update the minimum payout a trader can receive to be 5% of the trade amount.
process
set the minimum payout at mediation to be of trade amount proposal currently the minimum payout at mediation is this proposal suggests to increase this amount to be of trade amount reasoning with the introduction there needs to be a bigger incentive for the penalized party to accept the mediation proposal to prevent the trade entering arbitration this is because if the trade enters arbitration it will cause a delayed payout transaction and result in the penalized party likely not getting any reimbursement and the non penalized party likely not getting any compensation as a result outcome make the penalized party more likely to accept mediation by incentivizing them to receive a minimum of of the trade amount currently twice the minimum they can receive solution update the minimum payout a trader can receive to be of the trade amount
1
265,642
20,105,886,153
IssuesEvent
2022-02-07 10:26:22
Katolus/functions
https://api.github.com/repos/Katolus/functions
closed
Add support environments document
documentation
It should include information about support environments and runtimes.
1.0
Add support environments document - It should include information about support environments and runtimes.
non_process
add support environments document it should include information about support environments and runtimes
0
8,931
12,041,150,699
IssuesEvent
2020-04-14 08:20:32
Inria-Visages/BIDS-prov
https://api.github.com/repos/Inria-Visages/BIDS-prov
closed
List all features that are in the examples but not yet in the specification [0.5D].
Spec update process
Delivrable: Itemized list in a markdown document in [current repository](https://github.com/Inria-Visages/BIDS-prov/).
1.0
List all features that are in the examples but not yet in the specification [0.5D]. - Delivrable: Itemized list in a markdown document in [current repository](https://github.com/Inria-Visages/BIDS-prov/).
process
list all features that are in the examples but not yet in the specification delivrable itemized list in a markdown document in
1
5,389
8,213,270,862
IssuesEvent
2018-09-04 18:59:52
PennyDreadfulMTG/perf-reports
https://api.github.com/repos/PennyDreadfulMTG/perf-reports
closed
500 error at /api/gitpull
CalledProcessError decksite wontfix
Command '['pip', 'install', '-U', '--user', '-r', 'requirements.txt', '--no-cache']' returned non-zero exit status 1. Reported on decksite by logged_out -------------------------------------------------------------------------------- Request Method: POST Path: /api/gitpull? Cookies: {} Endpoint: process_github_webhook View Args: {} Person: logged_out Referrer: None Request Data: {} Host: pennydreadfulmagic.com Accept-Encoding: gzip Cf-Ipcountry: US X-Forwarded-For: 192.30.252.44, 162.158.106.14 Cf-Ray: 432a4fe03d7e797f-SEA X-Forwarded-Proto: https Cf-Visitor: {"scheme":"https"} Accept: */* User-Agent: GitHub-Hookshot/22e0d92 X-Github-Event: push X-Github-Delivery: c5c250be-7bc6-11e8-95c0-717918d50023 Content-Type: application/json Cf-Connecting-Ip: 192.30.252.44 X-Forwarded-Host: pennydreadfulmagic.com X-Forwarded-Server: pennydreadfulmagic.com Connection: Keep-Alive Content-Length: 9899-------------------------------------------------------------------------------- CalledProcessError Stack Trace: File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 2309, in __call__ return self.wsgi_app(environ, start_response) File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 2295, in wsgi_app response = self.handle_exception(e) File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 2292, in wsgi_app response = self.full_dispatch_request() File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1815, in full_dispatch_request rv = self.handle_user_exception(e) File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1718, in handle_user_exception reraise(exc_type, exc_value, tb) File "/home/discord/.local/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise raise value File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1813, in full_dispatch_request rv = self.dispatch_request() File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1799, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "./shared_web/api.py", line 19, in process_github_webhook subprocess.check_output(['pip', 'install', '-U', '--user', '-r', 'requirements.txt', '--no-cache']) File "/usr/lib64/python3.6/subprocess.py", line 336, in check_output **kwargs).stdout File "/usr/lib64/python3.6/subprocess.py", line 418, in run output=stdout, stderr=stderr)
1.0
500 error at /api/gitpull - Command '['pip', 'install', '-U', '--user', '-r', 'requirements.txt', '--no-cache']' returned non-zero exit status 1. Reported on decksite by logged_out -------------------------------------------------------------------------------- Request Method: POST Path: /api/gitpull? Cookies: {} Endpoint: process_github_webhook View Args: {} Person: logged_out Referrer: None Request Data: {} Host: pennydreadfulmagic.com Accept-Encoding: gzip Cf-Ipcountry: US X-Forwarded-For: 192.30.252.44, 162.158.106.14 Cf-Ray: 432a4fe03d7e797f-SEA X-Forwarded-Proto: https Cf-Visitor: {"scheme":"https"} Accept: */* User-Agent: GitHub-Hookshot/22e0d92 X-Github-Event: push X-Github-Delivery: c5c250be-7bc6-11e8-95c0-717918d50023 Content-Type: application/json Cf-Connecting-Ip: 192.30.252.44 X-Forwarded-Host: pennydreadfulmagic.com X-Forwarded-Server: pennydreadfulmagic.com Connection: Keep-Alive Content-Length: 9899-------------------------------------------------------------------------------- CalledProcessError Stack Trace: File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 2309, in __call__ return self.wsgi_app(environ, start_response) File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 2295, in wsgi_app response = self.handle_exception(e) File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 2292, in wsgi_app response = self.full_dispatch_request() File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1815, in full_dispatch_request rv = self.handle_user_exception(e) File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1718, in handle_user_exception reraise(exc_type, exc_value, tb) File "/home/discord/.local/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise raise value File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1813, in full_dispatch_request rv = self.dispatch_request() File "/home/discord/.local/lib/python3.6/site-packages/flask/app.py", line 1799, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "./shared_web/api.py", line 19, in process_github_webhook subprocess.check_output(['pip', 'install', '-U', '--user', '-r', 'requirements.txt', '--no-cache']) File "/usr/lib64/python3.6/subprocess.py", line 336, in check_output **kwargs).stdout File "/usr/lib64/python3.6/subprocess.py", line 418, in run output=stdout, stderr=stderr)
process
error at api gitpull command returned non zero exit status reported on decksite by logged out request method post path api gitpull cookies endpoint process github webhook view args person logged out referrer none request data host pennydreadfulmagic com accept encoding gzip cf ipcountry us x forwarded for cf ray sea x forwarded proto https cf visitor scheme https accept user agent github hookshot x github event push x github delivery content type application json cf connecting ip x forwarded host pennydreadfulmagic com x forwarded server pennydreadfulmagic com connection keep alive content length calledprocesserror stack trace file home discord local lib site packages flask app py line in call return self wsgi app environ start response file home discord local lib site packages flask app py line in wsgi app response self handle exception e file home discord local lib site packages flask app py line in wsgi app response self full dispatch request file home discord local lib site packages flask app py line in full dispatch request rv self handle user exception e file home discord local lib site packages flask app py line in handle user exception reraise exc type exc value tb file home discord local lib site packages flask compat py line in reraise raise value file home discord local lib site packages flask app py line in full dispatch request rv self dispatch request file home discord local lib site packages flask app py line in dispatch request return self view functions req view args file shared web api py line in process github webhook subprocess check output file usr subprocess py line in check output kwargs stdout file usr subprocess py line in run output stdout stderr stderr
1
333,047
10,114,645,889
IssuesEvent
2019-07-30 19:42:53
rstudio/gt
https://api.github.com/repos/rstudio/gt
closed
cols_label overrides tab_styles settings
Difficulty: ② Intermediate Effort: ② Medium Priority: ♨︎ Critical Type: ☹︎ Bug
If you use the cols_label function the formatting of the column header cells seems to be overridden by that function such that you can no longer control the formatting with tab_style. For example, this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = TRUE) ) ``` yields this ![Rplot1](https://user-images.githubusercontent.com/46826199/60226012-38367c80-983e-11e9-9b4b-70311cc0102c.png) Whereas this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::cols_label(hp = 'Horsepower') %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = TRUE) ) ``` yields this ![Rplot2](https://user-images.githubusercontent.com/46826199/60226068-7469dd00-983e-11e9-9105-0d36c0b86d4f.png) Ordering of the functions and other ways of trying to select the column header for styling don't seem to make a difference, so this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = gt::vars(hp)) ) ``` yields this ![Rplot3](https://user-images.githubusercontent.com/46826199/60226159-d4608380-983e-11e9-8a30-72d9bd6be8e0.png) But this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = gt::vars(hp)) ) %>% gt::cols_label(hp = 'Horsepower') ``` yields this ![Rplot4](https://user-images.githubusercontent.com/46826199/60226180-eb06da80-983e-11e9-8a12-dd706937d459.png) I know you can include markup in the cols_label function, so I'm possibly just using these functions incorrectly... but I ended up just working around this by relabeling my columns with dplyr before funneling them to gt() so that I wouldn't have to do manual formatting of the column headers in markdown or html.
1.0
cols_label overrides tab_styles settings - If you use the cols_label function the formatting of the column header cells seems to be overridden by that function such that you can no longer control the formatting with tab_style. For example, this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = TRUE) ) ``` yields this ![Rplot1](https://user-images.githubusercontent.com/46826199/60226012-38367c80-983e-11e9-9b4b-70311cc0102c.png) Whereas this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::cols_label(hp = 'Horsepower') %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = TRUE) ) ``` yields this ![Rplot2](https://user-images.githubusercontent.com/46826199/60226068-7469dd00-983e-11e9-9105-0d36c0b86d4f.png) Ordering of the functions and other ways of trying to select the column header for styling don't seem to make a difference, so this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = gt::vars(hp)) ) ``` yields this ![Rplot3](https://user-images.githubusercontent.com/46826199/60226159-d4608380-983e-11e9-8a30-72d9bd6be8e0.png) But this ``` gt::gtcars %>% dplyr::select(mfr, model, year, hp) %>% dplyr::top_n(n = 10) %>% gt::gt() %>% gt::tab_style( style = gt::cells_styles(text_size = '30px'), locations = gt::cells_column_labels(columns = gt::vars(hp)) ) %>% gt::cols_label(hp = 'Horsepower') ``` yields this ![Rplot4](https://user-images.githubusercontent.com/46826199/60226180-eb06da80-983e-11e9-8a12-dd706937d459.png) I know you can include markup in the cols_label function, so I'm possibly just using these functions incorrectly... but I ended up just working around this by relabeling my columns with dplyr before funneling them to gt() so that I wouldn't have to do manual formatting of the column headers in markdown or html.
non_process
cols label overrides tab styles settings if you use the cols label function the formatting of the column header cells seems to be overridden by that function such that you can no longer control the formatting with tab style for example this gt gtcars dplyr select mfr model year hp dplyr top n n gt gt gt tab style style gt cells styles text size locations gt cells column labels columns true yields this whereas this gt gtcars dplyr select mfr model year hp dplyr top n n gt gt gt cols label hp horsepower gt tab style style gt cells styles text size locations gt cells column labels columns true yields this ordering of the functions and other ways of trying to select the column header for styling don t seem to make a difference so this gt gtcars dplyr select mfr model year hp dplyr top n n gt gt gt tab style style gt cells styles text size locations gt cells column labels columns gt vars hp yields this but this gt gtcars dplyr select mfr model year hp dplyr top n n gt gt gt tab style style gt cells styles text size locations gt cells column labels columns gt vars hp gt cols label hp horsepower yields this i know you can include markup in the cols label function so i m possibly just using these functions incorrectly but i ended up just working around this by relabeling my columns with dplyr before funneling them to gt so that i wouldn t have to do manual formatting of the column headers in markdown or html
0
196,542
14,878,206,924
IssuesEvent
2021-01-20 05:10:26
felipexpert1996/biblioteca-django-angular
https://api.github.com/repos/felipexpert1996/biblioteca-django-angular
closed
Criar testes unitários para api de crud de categoria de livros
back-end test
- [ ] Criar teste unitário para api de cadastro de categoria de livros; - [ ] Criar teste unitário para api de atualização de categoria de livros; - [ ] Criar teste unitário para api de deleção de categoria de livros; - [ ] Criar teste unitário para api de listagem de categoria de livros;
1.0
Criar testes unitários para api de crud de categoria de livros - - [ ] Criar teste unitário para api de cadastro de categoria de livros; - [ ] Criar teste unitário para api de atualização de categoria de livros; - [ ] Criar teste unitário para api de deleção de categoria de livros; - [ ] Criar teste unitário para api de listagem de categoria de livros;
non_process
criar testes unitários para api de crud de categoria de livros criar teste unitário para api de cadastro de categoria de livros criar teste unitário para api de atualização de categoria de livros criar teste unitário para api de deleção de categoria de livros criar teste unitário para api de listagem de categoria de livros
0
17,494
23,305,507,970
IssuesEvent
2022-08-07 23:50:04
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add [Reality Bites]
suggested title in process
Please add as much of the following info as you can: Title: Reality Bites Type (film/tv show): show Film or show in which it appears: Reality Bites Is the parent film/show streaming anywhere? About when in the parent film/show does it appear? Toward the end when Ben Stillers character shows Winona Ryder’s character the new show from her footage. Actual footage of the film/show can be seen (yes/no)? Yes through the whole film.
1.0
Add [Reality Bites] - Please add as much of the following info as you can: Title: Reality Bites Type (film/tv show): show Film or show in which it appears: Reality Bites Is the parent film/show streaming anywhere? About when in the parent film/show does it appear? Toward the end when Ben Stillers character shows Winona Ryder’s character the new show from her footage. Actual footage of the film/show can be seen (yes/no)? Yes through the whole film.
process
add please add as much of the following info as you can title reality bites type film tv show show film or show in which it appears reality bites is the parent film show streaming anywhere about when in the parent film show does it appear toward the end when ben stillers character shows winona ryder’s character the new show from her footage actual footage of the film show can be seen yes no yes through the whole film
1
6,255
9,215,603,022
IssuesEvent
2019-03-11 04:05:33
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
UseShellExecute on macOS passes Arguments to `open` instead of specified `filename`
area-System.Diagnostics.Process bug os-mac-os-x
Using PSCore6: ```powershell $si = [System.Diagnostics.ProcessStartInfo]::new() $si.UseShellExecute = $true $si.FileName = (Get-Command pwsh).Source $si.Arguments = "-c 1+1;read-host" [System.Diagnostics.Process]::Start($si) ``` Expected: Start new instance of PowerShell outputting answer of `2` and waiting for `Enter`, then exit. Actual: Help for `open` command is displayed since `-c` is not valid. When using `open`, arguments to the executable are passed as `--args <arguments>` The above script works as expected on Windows.
1.0
UseShellExecute on macOS passes Arguments to `open` instead of specified `filename` - Using PSCore6: ```powershell $si = [System.Diagnostics.ProcessStartInfo]::new() $si.UseShellExecute = $true $si.FileName = (Get-Command pwsh).Source $si.Arguments = "-c 1+1;read-host" [System.Diagnostics.Process]::Start($si) ``` Expected: Start new instance of PowerShell outputting answer of `2` and waiting for `Enter`, then exit. Actual: Help for `open` command is displayed since `-c` is not valid. When using `open`, arguments to the executable are passed as `--args <arguments>` The above script works as expected on Windows.
process
useshellexecute on macos passes arguments to open instead of specified filename using powershell si new si useshellexecute true si filename get command pwsh source si arguments c read host start si expected start new instance of powershell outputting answer of and waiting for enter then exit actual help for open command is displayed since c is not valid when using open arguments to the executable are passed as args the above script works as expected on windows
1
6,817
9,959,692,195
IssuesEvent
2019-07-06 09:24:59
artisan-roaster-scope/artisan
https://api.github.com/repos/artisan-roaster-scope/artisan
closed
Overwrite batch counter from settings file; No.... Does not overwrite batch prefix.
in process
## Expected Behavior Batch prefix may continue to be loaded from settings file, even if we choose not to overwrite batch counter from the settings file. ![Overwrite Batch Counter dialogue](https://user-images.githubusercontent.com/22232379/60566238-db384a80-9d98-11e9-8589-50bae1bc4341.jpg) ## Actual Behavior If we say 'NO' to 'Overwrite your current batch counter h2839 by c2749 from the settings file?' then neither the batch (2839) nor prefix (h) are overwritten. That seems reasonable and logical according to the question, yet may not be practical if the user uses the prefix to record which settings file/type of roast is being produced in this session. The user may be using multiple settings files, yet still wishes to keep each batch number sequential and unique irrespective of type of coffee roasted. I think it makes sense that the prefix is always loaded as this is a static setting of the settings file that, along with the other static settings, is chosen by the user. Unlike the batch number which is expected to change dynamically with roasting. N.B. For background, in my use case, I have 2 settings files. The main difference is in the Axes settings, one has higher time and temperature for a hotter, longer, dark roast (2nd crack). The other setting file is for 3rd wave lighter roasts. But also some of my settings for over-temperature alarms and aspects of roast automation are different. The, recently added, display of batch number during roasting has the prefix and is great reassurance at the start of roasting that we have remembered to open the correct settings. ## Specifications - Artisan Version: 2.0.0 - Artisan Build (number in brackets shown in the about box): a0a1face - Platform (Mac/Windows/Linux + OS version): W10 - Connected devices or roasting machine: Phidgets and Modbus
1.0
Overwrite batch counter from settings file; No.... Does not overwrite batch prefix. - ## Expected Behavior Batch prefix may continue to be loaded from settings file, even if we choose not to overwrite batch counter from the settings file. ![Overwrite Batch Counter dialogue](https://user-images.githubusercontent.com/22232379/60566238-db384a80-9d98-11e9-8589-50bae1bc4341.jpg) ## Actual Behavior If we say 'NO' to 'Overwrite your current batch counter h2839 by c2749 from the settings file?' then neither the batch (2839) nor prefix (h) are overwritten. That seems reasonable and logical according to the question, yet may not be practical if the user uses the prefix to record which settings file/type of roast is being produced in this session. The user may be using multiple settings files, yet still wishes to keep each batch number sequential and unique irrespective of type of coffee roasted. I think it makes sense that the prefix is always loaded as this is a static setting of the settings file that, along with the other static settings, is chosen by the user. Unlike the batch number which is expected to change dynamically with roasting. N.B. For background, in my use case, I have 2 settings files. The main difference is in the Axes settings, one has higher time and temperature for a hotter, longer, dark roast (2nd crack). The other setting file is for 3rd wave lighter roasts. But also some of my settings for over-temperature alarms and aspects of roast automation are different. The, recently added, display of batch number during roasting has the prefix and is great reassurance at the start of roasting that we have remembered to open the correct settings. ## Specifications - Artisan Version: 2.0.0 - Artisan Build (number in brackets shown in the about box): a0a1face - Platform (Mac/Windows/Linux + OS version): W10 - Connected devices or roasting machine: Phidgets and Modbus
process
overwrite batch counter from settings file no does not overwrite batch prefix expected behavior batch prefix may continue to be loaded from settings file even if we choose not to overwrite batch counter from the settings file actual behavior if we say no to overwrite your current batch counter by from the settings file then neither the batch nor prefix h are overwritten that seems reasonable and logical according to the question yet may not be practical if the user uses the prefix to record which settings file type of roast is being produced in this session the user may be using multiple settings files yet still wishes to keep each batch number sequential and unique irrespective of type of coffee roasted i think it makes sense that the prefix is always loaded as this is a static setting of the settings file that along with the other static settings is chosen by the user unlike the batch number which is expected to change dynamically with roasting n b for background in my use case i have settings files the main difference is in the axes settings one has higher time and temperature for a hotter longer dark roast crack the other setting file is for wave lighter roasts but also some of my settings for over temperature alarms and aspects of roast automation are different the recently added display of batch number during roasting has the prefix and is great reassurance at the start of roasting that we have remembered to open the correct settings specifications artisan version artisan build number in brackets shown in the about box platform mac windows linux os version connected devices or roasting machine phidgets and modbus
1
11,786
7,704,321,228
IssuesEvent
2018-05-21 11:48:55
JuliaLang/julia
https://api.github.com/repos/JuliaLang/julia
closed
code_warntype is significantly slower than on 0.6
performance regression
0.6: ``` julia> @time @code_warntype 1+1; 1.695684 seconds (522.70 k allocations: 27.840 MiB, 8.88% gc time) ``` 0.7: ``` julia> @time @code_warntype 1+1 5.536148 seconds (2.12 M allocations: 121.450 MiB, 1.51% gc time) ``` Profiling shows a bunch of type inference stuff.
True
code_warntype is significantly slower than on 0.6 - 0.6: ``` julia> @time @code_warntype 1+1; 1.695684 seconds (522.70 k allocations: 27.840 MiB, 8.88% gc time) ``` 0.7: ``` julia> @time @code_warntype 1+1 5.536148 seconds (2.12 M allocations: 121.450 MiB, 1.51% gc time) ``` Profiling shows a bunch of type inference stuff.
non_process
code warntype is significantly slower than on julia time code warntype seconds k allocations mib gc time julia time code warntype seconds m allocations mib gc time profiling shows a bunch of type inference stuff
0
16,754
21,922,071,612
IssuesEvent
2022-05-22 18:21:09
googleapis/gapic-generator-python
https://api.github.com/repos/googleapis/gapic-generator-python
closed
Snippetgen: PRs are not prefixed with type docs
type: process
When a snippetgen updates are rolled out, since the types are prefixed with `chore` it does not triggers a release, even though a release is normally required for a docs update. For example: this PR https://github.com/googleapis/python-scheduler/pull/233 updated problematic snippets, however a docs refresh was not made due to no release being out. Is it possible to have a `docs` type prefixed for these updates, so that the docs are updated in time?
1.0
Snippetgen: PRs are not prefixed with type docs - When a snippetgen updates are rolled out, since the types are prefixed with `chore` it does not triggers a release, even though a release is normally required for a docs update. For example: this PR https://github.com/googleapis/python-scheduler/pull/233 updated problematic snippets, however a docs refresh was not made due to no release being out. Is it possible to have a `docs` type prefixed for these updates, so that the docs are updated in time?
process
snippetgen prs are not prefixed with type docs when a snippetgen updates are rolled out since the types are prefixed with chore it does not triggers a release even though a release is normally required for a docs update for example this pr updated problematic snippets however a docs refresh was not made due to no release being out is it possible to have a docs type prefixed for these updates so that the docs are updated in time
1
5,148
7,928,260,214
IssuesEvent
2018-07-06 10:57:38
Open-EO/openeo-api
https://api.github.com/repos/Open-EO/openeo-api
opened
Change the definition of a process_graph
process graphs vote
Issues such as #89 or #106 require a change in the process graph definition as objects are required to/should be used. Values of a process argument are defined as follows: `<Value> := <string|number|array|boolean|null|Process>` Source: https://open-eo.github.io/openeo-api/v/0.3.0/processgraphs/index.html For the mentioned issues (and probably others) we probably need to change the definition of values to: `<Value> := <string|number|array|boolean|null|object|Process>`
1.0
Change the definition of a process_graph - Issues such as #89 or #106 require a change in the process graph definition as objects are required to/should be used. Values of a process argument are defined as follows: `<Value> := <string|number|array|boolean|null|Process>` Source: https://open-eo.github.io/openeo-api/v/0.3.0/processgraphs/index.html For the mentioned issues (and probably others) we probably need to change the definition of values to: `<Value> := <string|number|array|boolean|null|object|Process>`
process
change the definition of a process graph issues such as or require a change in the process graph definition as objects are required to should be used values of a process argument are defined as follows source for the mentioned issues and probably others we probably need to change the definition of values to
1
254,707
27,413,650,073
IssuesEvent
2023-03-01 12:19:34
scm-automation-project/npm-7-with-workspaces-without-lock-file-project
https://api.github.com/repos/scm-automation-project/npm-7-with-workspaces-without-lock-file-project
closed
package-b-1.0.0.tgz: 1 vulnerabilities (highest severity is: 5.0) - autoclosed
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>package-b-1.0.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/express/node_modules/mime/package.json,/node_modules/serve-static/node_modules/mime/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/npm-7-with-workspaces-without-lock-file-project/commit/6c8f09787e3c71d72460de2f7571429d885ffb9b">6c8f09787e3c71d72460de2f7571429d885ffb9b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (package-b version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2017-16138](https://www.mend.io/vulnerability-database/CVE-2017-16138) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | mime-1.3.4.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2017-16138</summary> ### Vulnerable Library - <b>mime-1.3.4.tgz</b></p> <p>A comprehensive library for mime-type mapping</p> <p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/express/node_modules/mime/package.json,/node_modules/serve-static/node_modules/mime/package.json</p> <p> Dependency Hierarchy: - package-b-1.0.0.tgz (Root Library) - express-4.15.5.tgz - send-0.15.6.tgz - :x: **mime-1.3.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/npm-7-with-workspaces-without-lock-file-project/commit/6c8f09787e3c71d72460de2f7571429d885ffb9b">6c8f09787e3c71d72460de2f7571429d885ffb9b</a></p> </p> <p></p> ### Vulnerability Details <p> The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-16138>CVE-2017-16138</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.0</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 1.4.1,2.0.3</p> </p> <p></p> </details>
True
package-b-1.0.0.tgz: 1 vulnerabilities (highest severity is: 5.0) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>package-b-1.0.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/express/node_modules/mime/package.json,/node_modules/serve-static/node_modules/mime/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/npm-7-with-workspaces-without-lock-file-project/commit/6c8f09787e3c71d72460de2f7571429d885ffb9b">6c8f09787e3c71d72460de2f7571429d885ffb9b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (package-b version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2017-16138](https://www.mend.io/vulnerability-database/CVE-2017-16138) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | mime-1.3.4.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2017-16138</summary> ### Vulnerable Library - <b>mime-1.3.4.tgz</b></p> <p>A comprehensive library for mime-type mapping</p> <p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/express/node_modules/mime/package.json,/node_modules/serve-static/node_modules/mime/package.json</p> <p> Dependency Hierarchy: - package-b-1.0.0.tgz (Root Library) - express-4.15.5.tgz - send-0.15.6.tgz - :x: **mime-1.3.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/scm-automation-project/npm-7-with-workspaces-without-lock-file-project/commit/6c8f09787e3c71d72460de2f7571429d885ffb9b">6c8f09787e3c71d72460de2f7571429d885ffb9b</a></p> </p> <p></p> ### Vulnerability Details <p> The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-16138>CVE-2017-16138</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.0</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 1.4.1,2.0.3</p> </p> <p></p> </details>
non_process
package b tgz vulnerabilities highest severity is autoclosed vulnerable library package b tgz path to dependency file package json path to vulnerable library node modules express node modules mime package json node modules serve static node modules mime package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in package b version remediation available medium mime tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library mime tgz a comprehensive library for mime type mapping library home page a href path to dependency file package json path to vulnerable library node modules express node modules mime package json node modules serve static node modules mime package json dependency hierarchy package b tgz root library express tgz send tgz x mime tgz vulnerable library found in head commit a href vulnerability details the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
11,873
14,674,174,395
IssuesEvent
2020-12-30 14:46:20
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Mobile] Closed study > Not able to enroll into a study
Bug P1 Participant manager Process: Tested dev
Closed study > Not able to enroll into a study Steps 1. Enroll into a closed study and withdrawn from the study 2. Add the withdrawn participant to another site present in the same study and invite 3. Try to enroll into the study AR : User is getting error message ER : User should be able to enroll into a study ![closed study1](https://user-images.githubusercontent.com/71445210/102502882-28030480-40a5-11eb-8de7-96b3d788015a.png) Issue observed in QA and Dev
1.0
[Mobile] Closed study > Not able to enroll into a study - Closed study > Not able to enroll into a study Steps 1. Enroll into a closed study and withdrawn from the study 2. Add the withdrawn participant to another site present in the same study and invite 3. Try to enroll into the study AR : User is getting error message ER : User should be able to enroll into a study ![closed study1](https://user-images.githubusercontent.com/71445210/102502882-28030480-40a5-11eb-8de7-96b3d788015a.png) Issue observed in QA and Dev
process
closed study not able to enroll into a study closed study not able to enroll into a study steps enroll into a closed study and withdrawn from the study add the withdrawn participant to another site present in the same study and invite try to enroll into the study ar user is getting error message er user should be able to enroll into a study issue observed in qa and dev
1
18,585
24,567,811,763
IssuesEvent
2022-10-13 05:49:04
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Android] Getting updated consent flow again in the following scenario
Bug P1 Android Process: Fixed Process: Tested QA Process: Tested dev
**Steps:** 1. Sign in / Sign up 2. Enroll to the study 3. Now, Go back to SB and update the consent flow again for enrolled participants 4. Open the mobile app 5. Complete the consent flow 6. Minimize the app after navigating to study activities screen and again open the app 7. Click on 'Home' icon 8. Again click on that particular study and Verify **AR:** Getting updated consent flow again in the following scenario **ER:** Updated consent flow should not be available again
3.0
[Android] Getting updated consent flow again in the following scenario - **Steps:** 1. Sign in / Sign up 2. Enroll to the study 3. Now, Go back to SB and update the consent flow again for enrolled participants 4. Open the mobile app 5. Complete the consent flow 6. Minimize the app after navigating to study activities screen and again open the app 7. Click on 'Home' icon 8. Again click on that particular study and Verify **AR:** Getting updated consent flow again in the following scenario **ER:** Updated consent flow should not be available again
process
getting updated consent flow again in the following scenario steps sign in sign up enroll to the study now go back to sb and update the consent flow again for enrolled participants open the mobile app complete the consent flow minimize the app after navigating to study activities screen and again open the app click on home icon again click on that particular study and verify ar getting updated consent flow again in the following scenario er updated consent flow should not be available again
1
17,203
22,779,860,804
IssuesEvent
2022-07-08 18:23:20
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/groupbytraceprocessor] how to use groupbytraceprocessor
priority:p2 processor/groupbytrace
is there any new information for groupbytraceprocessor? Is it ready for use?
1.0
[processor/groupbytraceprocessor] how to use groupbytraceprocessor - is there any new information for groupbytraceprocessor? Is it ready for use?
process
how to use groupbytraceprocessor is there any new information for groupbytraceprocessor is it ready for use?
1
306,793
23,172,238,390
IssuesEvent
2022-07-30 22:31:20
open-ephys/plugin-GUI
https://api.github.com/repos/open-ephys/plugin-GUI
closed
Computer specifications
Documentation
Hi all, First of all, Thanks a lot in advance for all the effort put in this project, I'm trying to get settled with open ephys and would like some advice on the computer hardware and system requirements for performing recordings from >=128 simultaneous channels , with potential online PSTH computation and sorting . Are there any minimal requirements in the RAM size, drive size , processor (in windows) for preventing crashes while recording? thanks a lot! sandra
1.0
Computer specifications - Hi all, First of all, Thanks a lot in advance for all the effort put in this project, I'm trying to get settled with open ephys and would like some advice on the computer hardware and system requirements for performing recordings from >=128 simultaneous channels , with potential online PSTH computation and sorting . Are there any minimal requirements in the RAM size, drive size , processor (in windows) for preventing crashes while recording? thanks a lot! sandra
non_process
computer specifications hi all first of all thanks a lot in advance for all the effort put in this project i m trying to get settled with open ephys and would like some advice on the computer hardware and system requirements for performing recordings from simultaneous channels with potential online psth computation and sorting are there any minimal requirements in the ram size drive size processor in windows for preventing crashes while recording thanks a lot sandra
0
84,726
24,398,436,163
IssuesEvent
2022-10-04 21:43:23
opensearch-project/OpenSearch-Dashboards
https://api.github.com/repos/opensearch-project/OpenSearch-Dashboards
closed
[VizBuilder] Create a new wizard directly on dashboard
enhancement vis builder
Subtask of https://github.com/opensearch-project/OpenSearch-Dashboards/issues/960 Requirements: * Be able to create a new wizard directly from the dashboard by clicking 'Edit' then 'Create new' * The save modal shows option to toggle between add to dashboard after saving or not <img width="513" alt="Screen Shot 2022-09-16 at 4 00 49 PM" src="https://user-images.githubusercontent.com/43937633/190828431-60b585ed-d1fe-4180-bab5-2036a0f54e3d.png">
1.0
[VizBuilder] Create a new wizard directly on dashboard - Subtask of https://github.com/opensearch-project/OpenSearch-Dashboards/issues/960 Requirements: * Be able to create a new wizard directly from the dashboard by clicking 'Edit' then 'Create new' * The save modal shows option to toggle between add to dashboard after saving or not <img width="513" alt="Screen Shot 2022-09-16 at 4 00 49 PM" src="https://user-images.githubusercontent.com/43937633/190828431-60b585ed-d1fe-4180-bab5-2036a0f54e3d.png">
non_process
create a new wizard directly on dashboard subtask of requirements be able to create a new wizard directly from the dashboard by clicking edit then create new the save modal shows option to toggle between add to dashboard after saving or not img width alt screen shot at pm src
0
44,281
13,051,274,229
IssuesEvent
2020-07-29 16:49:09
jtimberlake/csp-html-webpack-plugin
https://api.github.com/repos/jtimberlake/csp-html-webpack-plugin
opened
CVE-2020-13822 (High) detected in elliptic-6.5.1.tgz
security vulnerability
## CVE-2020-13822 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.1.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/csp-html-webpack-plugin/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/csp-html-webpack-plugin/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - webpack-4.41.2.tgz (Root Library) - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/csp-html-webpack-plugin/commit/e90e35b4f77c718e61c7abd9b80bee6db5dc4fdd">e90e35b4f77c718e61c7abd9b80bee6db5dc4fdd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Elliptic package 6.5.2 for Node.js allows ECDSA signature malleability via variations in encoding, leading '\0' bytes, or integer overflows. This could conceivably have a security-relevant impact if an application relied on a single canonical signature. <p>Publish Date: 2020-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13822>CVE-2020-13822</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/indutny/elliptic/tree/v6.5.3">https://github.com/indutny/elliptic/tree/v6.5.3</a></p> <p>Release Date: 2020-06-04</p> <p>Fix Resolution: v6.5.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.5.1","isTransitiveDependency":true,"dependencyTree":"webpack:4.41.2;node-libs-browser:2.2.1;crypto-browserify:3.12.0;browserify-sign:4.0.4;elliptic:6.5.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.3"}],"vulnerabilityIdentifier":"CVE-2020-13822","vulnerabilityDetails":"The Elliptic package 6.5.2 for Node.js allows ECDSA signature malleability via variations in encoding, leading \u0027\\0\u0027 bytes, or integer overflows. This could conceivably have a security-relevant impact if an application relied on a single canonical signature.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13822","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-13822 (High) detected in elliptic-6.5.1.tgz - ## CVE-2020-13822 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.1.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/csp-html-webpack-plugin/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/csp-html-webpack-plugin/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - webpack-4.41.2.tgz (Root Library) - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/csp-html-webpack-plugin/commit/e90e35b4f77c718e61c7abd9b80bee6db5dc4fdd">e90e35b4f77c718e61c7abd9b80bee6db5dc4fdd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Elliptic package 6.5.2 for Node.js allows ECDSA signature malleability via variations in encoding, leading '\0' bytes, or integer overflows. This could conceivably have a security-relevant impact if an application relied on a single canonical signature. <p>Publish Date: 2020-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13822>CVE-2020-13822</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/indutny/elliptic/tree/v6.5.3">https://github.com/indutny/elliptic/tree/v6.5.3</a></p> <p>Release Date: 2020-06-04</p> <p>Fix Resolution: v6.5.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.5.1","isTransitiveDependency":true,"dependencyTree":"webpack:4.41.2;node-libs-browser:2.2.1;crypto-browserify:3.12.0;browserify-sign:4.0.4;elliptic:6.5.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.3"}],"vulnerabilityIdentifier":"CVE-2020-13822","vulnerabilityDetails":"The Elliptic package 6.5.2 for Node.js allows ECDSA signature malleability via variations in encoding, leading \u0027\\0\u0027 bytes, or integer overflows. This could conceivably have a security-relevant impact if an application relied on a single canonical signature.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13822","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in elliptic tgz cve high severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file tmp ws scm csp html webpack plugin package json path to vulnerable library tmp ws scm csp html webpack plugin node modules elliptic package json dependency hierarchy webpack tgz root library node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library found in head commit a href vulnerability details the elliptic package for node js allows ecdsa signature malleability via variations in encoding leading bytes or integer overflows this could conceivably have a security relevant impact if an application relied on a single canonical signature publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the elliptic package for node js allows ecdsa signature malleability via variations in encoding leading bytes or integer overflows this could conceivably have a security relevant impact if an application relied on a single canonical signature vulnerabilityurl
0
10,847
13,627,108,199
IssuesEvent
2020-09-24 12:06:52
rladies/rladiesguide
https://api.github.com/repos/rladies/rladiesguide
closed
Assess use of current-chapters.csv
rladies processes :bullettrain_side:
@yabellini @ledell if processes involving the CSV change I'll need to update quite a few sections cf https://github.com/rladies/rladiesguide/search?q=current-chapters.csv&unscoped_q=current-chapters.csv
1.0
Assess use of current-chapters.csv - @yabellini @ledell if processes involving the CSV change I'll need to update quite a few sections cf https://github.com/rladies/rladiesguide/search?q=current-chapters.csv&unscoped_q=current-chapters.csv
process
assess use of current chapters csv yabellini ledell if processes involving the csv change i ll need to update quite a few sections cf
1
299,918
9,205,970,030
IssuesEvent
2019-03-08 12:16:45
qissue-bot/QGIS
https://api.github.com/repos/qissue-bot/QGIS
closed
PostGIS: missing buttons to delete and add new columns in table editor
Category: GUI Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report
--- Author Name: **Maciej Sieczka -** (Maciej Sieczka -) Original Redmine Issue: 1273, https://issues.qgis.org/issues/1273 Original Assignee: nobody - --- Why are the buttons to delete and add new columns for [[PostGIS]] layers missing in the table editor? I can bet they were there few revisions ago.
1.0
PostGIS: missing buttons to delete and add new columns in table editor - --- Author Name: **Maciej Sieczka -** (Maciej Sieczka -) Original Redmine Issue: 1273, https://issues.qgis.org/issues/1273 Original Assignee: nobody - --- Why are the buttons to delete and add new columns for [[PostGIS]] layers missing in the table editor? I can bet they were there few revisions ago.
non_process
postgis missing buttons to delete and add new columns in table editor author name maciej sieczka maciej sieczka original redmine issue original assignee nobody why are the buttons to delete and add new columns for layers missing in the table editor i can bet they were there few revisions ago
0
20,766
27,500,829,779
IssuesEvent
2023-03-05 17:23:02
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
opened
Add NBN Atlas as a source
data processing back end new source
<!-- Please populate as much information as you can, but don't worry if you can't fill out all fields --> The National Biodiversity Network(NBN) has a portal listing 1000+ datasets on biodiversity and wildlife species in the UK [https://registry.nbnatlas.org/datasets](https://registry.nbnatlas.org/datasets) **Link to the data source** [https://api.nbnatlas.org](https://api.nbnatlas.org) or **Data source type (if known)** Looks like a custom API **Organization(s) the data belongs to** NBN Atlas is a multi-owner portal **What licences are applied to the data being published? (if known)** [https://docs.nbnatlas.org/guidance-for-using-data/](https://docs.nbnatlas.org/guidance-for-using-data/) CC0, CC-BY, CC-BY-NC, and OGL. **Additional information** It looks like the user portal is behind a login, it may be that the API requires user credentials to access - to investigate. This is also a UK portal so may want to see if we can filter down to Scotland level, or, datasets containing Scottish data. Or see [https://scotland.nbnatlas.org](https://scotland.nbnatlas.org)
1.0
Add NBN Atlas as a source - <!-- Please populate as much information as you can, but don't worry if you can't fill out all fields --> The National Biodiversity Network(NBN) has a portal listing 1000+ datasets on biodiversity and wildlife species in the UK [https://registry.nbnatlas.org/datasets](https://registry.nbnatlas.org/datasets) **Link to the data source** [https://api.nbnatlas.org](https://api.nbnatlas.org) or **Data source type (if known)** Looks like a custom API **Organization(s) the data belongs to** NBN Atlas is a multi-owner portal **What licences are applied to the data being published? (if known)** [https://docs.nbnatlas.org/guidance-for-using-data/](https://docs.nbnatlas.org/guidance-for-using-data/) CC0, CC-BY, CC-BY-NC, and OGL. **Additional information** It looks like the user portal is behind a login, it may be that the API requires user credentials to access - to investigate. This is also a UK portal so may want to see if we can filter down to Scotland level, or, datasets containing Scottish data. Or see [https://scotland.nbnatlas.org](https://scotland.nbnatlas.org)
process
add nbn atlas as a source the national biodiversity network nbn has a portal listing datasets on biodiversity and wildlife species in the uk link to the data source or data source type if known looks like a custom api organization s the data belongs to nbn atlas is a multi owner portal what licences are applied to the data being published if known cc by cc by nc and ogl additional information it looks like the user portal is behind a login it may be that the api requires user credentials to access to investigate this is also a uk portal so may want to see if we can filter down to scotland level or datasets containing scottish data or see
1
36,615
17,796,872,755
IssuesEvent
2021-08-31 23:59:23
microsoft/TypeScript
https://api.github.com/repos/microsoft/TypeScript
opened
Experiment with pre-allocated collections
Needs Investigation Domain: Performance
In some (many?) places, we know in advance how big the array we're building will be but, to keep it packed (vs sparse), we have to push elements one at a time and let the array grow in chunks, generating a bunch of garbage. We hypothesize that allocating the correct size up-front will save us a bunch of GC time. With maps, we less commonly know the final size, but we want to experiment with giving them capacity 1 (rather than the default, 4) in cases where this is the most common size. (Something about symbols? Maybe most aren't merged?) https://chromium-review.googlesource.com/c/v8/v8/+/3122568 https://chromium-review.googlesource.com/c/v8/v8/+/3122502 To play with this, we'll need to patch and build node locally.
True
Experiment with pre-allocated collections - In some (many?) places, we know in advance how big the array we're building will be but, to keep it packed (vs sparse), we have to push elements one at a time and let the array grow in chunks, generating a bunch of garbage. We hypothesize that allocating the correct size up-front will save us a bunch of GC time. With maps, we less commonly know the final size, but we want to experiment with giving them capacity 1 (rather than the default, 4) in cases where this is the most common size. (Something about symbols? Maybe most aren't merged?) https://chromium-review.googlesource.com/c/v8/v8/+/3122568 https://chromium-review.googlesource.com/c/v8/v8/+/3122502 To play with this, we'll need to patch and build node locally.
non_process
experiment with pre allocated collections in some many places we know in advance how big the array we re building will be but to keep it packed vs sparse we have to push elements one at a time and let the array grow in chunks generating a bunch of garbage we hypothesize that allocating the correct size up front will save us a bunch of gc time with maps we less commonly know the final size but we want to experiment with giving them capacity rather than the default in cases where this is the most common size something about symbols maybe most aren t merged to play with this we ll need to patch and build node locally
0
9,660
12,641,951,737
IssuesEvent
2020-06-16 07:17:33
microsoft/react-native-windows
https://api.github.com/repos/microsoft/react-native-windows
closed
Rename "master" Builds to "canary" or "nightly"
Area: Release Process
We currently release builds from our development branch as `0.0.0-master.xxx`. This scheme seems to be causing confusion. We should switch to `canary` to better set expectations. This should be safe to do since we haven't yet externally documented guarantees.
1.0
Rename "master" Builds to "canary" or "nightly" - We currently release builds from our development branch as `0.0.0-master.xxx`. This scheme seems to be causing confusion. We should switch to `canary` to better set expectations. This should be safe to do since we haven't yet externally documented guarantees.
process
rename master builds to canary or nightly we currently release builds from our development branch as master xxx this scheme seems to be causing confusion we should switch to canary to better set expectations this should be safe to do since we haven t yet externally documented guarantees
1
200,405
15,105,593,450
IssuesEvent
2021-02-08 13:15:42
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
Typo "Promethues" when installing v2 Monitoring in Apps & Marketplace
[zube]: To Test
<!-- Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase. --> **What kind of request is this (question/bug/enhancement/feature request): bug/typo** **Steps to reproduce (least amount of steps as possible):** * Go to Cluster Explorer, select `Apps & Marketplace` and click on Monitoring chart * Side-tabs pane for Monitoring settings contains entry **"Promethues"** with the typo + the same for the title of this section. **Result:** ![promethues_typo](https://user-images.githubusercontent.com/12828077/105496433-bdee0880-5cbd-11eb-8a39-66b48a077306.png) **Environment information** - Monitoring chart 9.4.202 - Rancher version: v2.5.6-rc2 - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below -->
1.0
Typo "Promethues" when installing v2 Monitoring in Apps & Marketplace - <!-- Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase. --> **What kind of request is this (question/bug/enhancement/feature request): bug/typo** **Steps to reproduce (least amount of steps as possible):** * Go to Cluster Explorer, select `Apps & Marketplace` and click on Monitoring chart * Side-tabs pane for Monitoring settings contains entry **"Promethues"** with the typo + the same for the title of this section. **Result:** ![promethues_typo](https://user-images.githubusercontent.com/12828077/105496433-bdee0880-5cbd-11eb-8a39-66b48a077306.png) **Environment information** - Monitoring chart 9.4.202 - Rancher version: v2.5.6-rc2 - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below -->
non_process
typo promethues when installing monitoring in apps marketplace please search for existing issues first then read to see what we expect in an issue for security issues please email security rancher com instead of posting a public issue in github you may but are not required to use the gpg key located on keybase what kind of request is this question bug enhancement feature request bug typo steps to reproduce least amount of steps as possible go to cluster explorer select apps marketplace and click on monitoring chart side tabs pane for monitoring settings contains entry promethues with the typo the same for the title of this section result environment information monitoring chart rancher version installation option single install ha single if the reported issue is regarding a created cluster please provide requested info below
0
13,131
15,530,420,653
IssuesEvent
2021-03-13 19:01:32
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
[FALSE-POSITIVE] aetna.com
whitelisting process
*@famouspotatoes commented on Dec 1, 2020, 5:11 PM UTC:* This is a safe website for health insurance *This issue was moved by [dnmTX](https://github.com/dnmTX) from [Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist#594](https://github.com/Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist/issues/594).*
1.0
[FALSE-POSITIVE] aetna.com - *@famouspotatoes commented on Dec 1, 2020, 5:11 PM UTC:* This is a safe website for health insurance *This issue was moved by [dnmTX](https://github.com/dnmTX) from [Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist#594](https://github.com/Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist/issues/594).*
process
aetna com famouspotatoes commented on dec pm utc this is a safe website for health insurance this issue was moved by from
1
11,536
14,410,026,798
IssuesEvent
2020-12-04 03:44:17
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Error when using bazel build in angular 8
type: support / not a bug (process)
I'm trying to build angular project with bazel build & getting below error. ERROR: /Users/ztna_userportal_applogo/gateway/ui/src/app/BUILD.bazel:37:10: Compiling Angular templates (ngc) //ui/src/app:app failed (Exit 1): ngc-wrapped.sh failed: error executing command (cd /private/var/tmp/_bazel_amit.katyal/c6d6c66a668f2db56ab82b64d688c088/execroot/ztna_gateway && \ exec env - \ bazel-out/host/bin/external/npm/@angular/bazel/bin/ngc-wrapped.sh '--node_options=--expose-gc' '--node_options=--stack-trace-limit=100' '--node_options=--max-old-space-size=4096' @@bazel-out/darwin-fastbuild/bin/ui/src/app/app_es5_tsconfig.json) Execution platform: @local_config_platform//:host ui/src/app/user-layout/user-layout.component.ts(37,10): error TS2742: T**he inferred type of 'transform' cannot be named without a reference to '../../../../external/npm/node_modules/@angular/platform-browser/platform-browser'. This is likely not portable. A type annotation is necessary.** Code changes import { Component, OnInit } from '@angular/core'; import { DomSanitizer } from '@angular/platform-browser'; import { ApplicationService } from '../service/application.service'; @Component({ selector: 'app-user-layout', templateUrl: './user-layout.component.html', styleUrls: ['./user-layout.component.scss'], }) **public transform(encodedString) { ==============> BUILD ERROR** return this.domSanitizer.bypassSecurityTrustResourceUrl('data:image/png;base64,' + encodedString); } Any idea what is the issue here ?
1.0
Error when using bazel build in angular 8 - I'm trying to build angular project with bazel build & getting below error. ERROR: /Users/ztna_userportal_applogo/gateway/ui/src/app/BUILD.bazel:37:10: Compiling Angular templates (ngc) //ui/src/app:app failed (Exit 1): ngc-wrapped.sh failed: error executing command (cd /private/var/tmp/_bazel_amit.katyal/c6d6c66a668f2db56ab82b64d688c088/execroot/ztna_gateway && \ exec env - \ bazel-out/host/bin/external/npm/@angular/bazel/bin/ngc-wrapped.sh '--node_options=--expose-gc' '--node_options=--stack-trace-limit=100' '--node_options=--max-old-space-size=4096' @@bazel-out/darwin-fastbuild/bin/ui/src/app/app_es5_tsconfig.json) Execution platform: @local_config_platform//:host ui/src/app/user-layout/user-layout.component.ts(37,10): error TS2742: T**he inferred type of 'transform' cannot be named without a reference to '../../../../external/npm/node_modules/@angular/platform-browser/platform-browser'. This is likely not portable. A type annotation is necessary.** Code changes import { Component, OnInit } from '@angular/core'; import { DomSanitizer } from '@angular/platform-browser'; import { ApplicationService } from '../service/application.service'; @Component({ selector: 'app-user-layout', templateUrl: './user-layout.component.html', styleUrls: ['./user-layout.component.scss'], }) **public transform(encodedString) { ==============> BUILD ERROR** return this.domSanitizer.bypassSecurityTrustResourceUrl('data:image/png;base64,' + encodedString); } Any idea what is the issue here ?
process
error when using bazel build in angular i m trying to build angular project with bazel build getting below error error users ztna userportal applogo gateway ui src app build bazel compiling angular templates ngc ui src app app failed exit ngc wrapped sh failed error executing command cd private var tmp bazel amit katyal execroot ztna gateway exec env bazel out host bin external npm angular bazel bin ngc wrapped sh node options expose gc node options stack trace limit node options max old space size bazel out darwin fastbuild bin ui src app app tsconfig json execution platform local config platform host ui src app user layout user layout component ts error t he inferred type of transform cannot be named without a reference to external npm node modules angular platform browser platform browser this is likely not portable a type annotation is necessary code changes import component oninit from angular core import domsanitizer from angular platform browser import applicationservice from service application service component selector app user layout templateurl user layout component html styleurls public transform encodedstring build error return this domsanitizer bypasssecuritytrustresourceurl data image png encodedstring any idea what is the issue here
1
203,274
23,133,447,685
IssuesEvent
2022-07-28 12:31:19
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[Security Solution]: On selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects.
bug triage_needed impact:medium Team: SecuritySolution v8.4.0
**Describe the bug:** On selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects. **Build Details:** ``` VERSION : 8.4.0-SNAPSHOT BUILD: 54932 COMMIT: 0645a3ba388f81fc6b43136e64ad53614101e947 ``` **Preconditions** 1. Kibana should be running. **Steps to Reproduce** 1. Navigate to Stack Management. 2. Now, navigate to Tags tab and create multiple tags with name `Security Solution`. 4. Now, add system dashboard to the created tags. 5. Go to Saved Objects tab. 6. Now, click on Tags selection dropdown and select on any tag created with same name i.e `Security Solution`. 7. Observe that on selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects. **Actual Result** On selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects. **Expected Result** On selecting a particular tag all available tags should not get select and unselect simultaneously with same name and correct tag assignments should show under Saved Objects. **Screen-Recording:** https://user-images.githubusercontent.com/84007952/181504210-cfa43640-3ce1-4442-9454-127a0b4ec97e.mp4
True
[Security Solution]: On selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects. - **Describe the bug:** On selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects. **Build Details:** ``` VERSION : 8.4.0-SNAPSHOT BUILD: 54932 COMMIT: 0645a3ba388f81fc6b43136e64ad53614101e947 ``` **Preconditions** 1. Kibana should be running. **Steps to Reproduce** 1. Navigate to Stack Management. 2. Now, navigate to Tags tab and create multiple tags with name `Security Solution`. 4. Now, add system dashboard to the created tags. 5. Go to Saved Objects tab. 6. Now, click on Tags selection dropdown and select on any tag created with same name i.e `Security Solution`. 7. Observe that on selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects. **Actual Result** On selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under Saved Objects. **Expected Result** On selecting a particular tag all available tags should not get select and unselect simultaneously with same name and correct tag assignments should show under Saved Objects. **Screen-Recording:** https://user-images.githubusercontent.com/84007952/181504210-cfa43640-3ce1-4442-9454-127a0b4ec97e.mp4
non_process
on selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under saved objects describe the bug on selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under saved objects build details version snapshot build commit preconditions kibana should be running steps to reproduce navigate to stack management now navigate to tags tab and create multiple tags with name security solution now add system dashboard to the created tags go to saved objects tab now click on tags selection dropdown and select on any tag created with same name i e security solution observe that on selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under saved objects actual result on selecting a particular tag all available tags gets selected and unselected simultaneously with same name and incorrect tag assignments shows under saved objects expected result on selecting a particular tag all available tags should not get select and unselect simultaneously with same name and correct tag assignments should show under saved objects screen recording
0
5,567
8,407,222,146
IssuesEvent
2018-10-11 20:17:15
googleapis/google-cloud-java
https://api.github.com/repos/googleapis/google-cloud-java
closed
Create GitHub Issue/PR templates
status: in progress type: process
This should help users file more detailed bug reports and aid us in reproducing and fixing bugs.
1.0
Create GitHub Issue/PR templates - This should help users file more detailed bug reports and aid us in reproducing and fixing bugs.
process
create github issue pr templates this should help users file more detailed bug reports and aid us in reproducing and fixing bugs
1
16,323
20,976,838,288
IssuesEvent
2022-03-28 15:55:56
googleapis/python-db-dtypes-pandas
https://api.github.com/repos/googleapis/python-db-dtypes-pandas
closed
release 1.0 package
type: process api: bigquery
Now that we are passing the full compliance test suite & testing against prerelease versions of pandas, I'm comfortable calling this a 1.0 package.
1.0
release 1.0 package - Now that we are passing the full compliance test suite & testing against prerelease versions of pandas, I'm comfortable calling this a 1.0 package.
process
release package now that we are passing the full compliance test suite testing against prerelease versions of pandas i m comfortable calling this a package
1
9,696
12,700,108,222
IssuesEvent
2020-06-22 15:49:06
prisma/vscode
https://api.github.com/repos/prisma/vscode
closed
Document "format on save" in READMEs
kind/improvement process/candidate
See https://github.com/prisma/vscode/issues/258 This should probably be highlighted in the docs as it is a high value change for users of the extension.
1.0
Document "format on save" in READMEs - See https://github.com/prisma/vscode/issues/258 This should probably be highlighted in the docs as it is a high value change for users of the extension.
process
document format on save in readmes see this should probably be highlighted in the docs as it is a high value change for users of the extension
1
22,429
31,160,649,054
IssuesEvent
2023-08-16 15:45:52
GoogleCloudPlatform/contact-center-ai-samples
https://api.github.com/repos/GoogleCloudPlatform/contact-center-ai-samples
closed
Dependency Dashboard
type: process
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Rate-Limited These updates are currently rate-limited. Click on a checkbox below to force their creation now. - [ ] <!-- unlimit-branch=renovate/prettier-2.x-lockfile -->chore(deps): update dependency prettier to v2.8.8 - [ ] <!-- unlimit-branch=renovate/stylelint-15.x-lockfile -->chore(deps): update dependency stylelint to v15.10.2 - [ ] <!-- unlimit-branch=renovate/eslint-plugin-react-7.x-lockfile -->chore(deps): update dependency eslint-plugin-react to v7.33.2 - [ ] <!-- unlimit-branch=renovate/tanstack-query-monorepo -->fix(deps): update dependency @tanstack/react-query to v4.32.6 - [ ] <!-- unlimit-branch=renovate/testing-library-jest-dom-5.x-lockfile -->fix(deps): update dependency @testing-library/jest-dom to v5.17.0 - [ ] <!-- unlimit-branch=renovate/axios-0.x -->fix(deps): update dependency axios to ^0.27.0 - [ ] <!-- unlimit-branch=renovate/react-router-monorepo -->fix(deps): update dependency react-router-dom to v6.15.0 - [ ] <!-- unlimit-branch=renovate/emotion-monorepo -->fix(deps): update emotion monorepo to v11.11.1 (`@emotion/react`, `@emotion/styled`) - [ ] <!-- unlimit-branch=renovate/material-ui-monorepo -->fix(deps): update material-ui monorepo (`@mui/icons-material`, `@mui/material`) - [ ] <!-- unlimit-branch=renovate/networkx-3.x -->chore(deps): update dependency networkx to v3 - [ ] <!-- unlimit-branch=renovate/prettier-3.x -->chore(deps): update dependency prettier to v3 - [ ] <!-- unlimit-branch=renovate/stylelint-config-standard-34.x -->chore(deps): update dependency stylelint-config-standard to v34 - [ ] <!-- unlimit-branch=renovate/svelte-4.x -->chore(deps): update dependency svelte to v4 - [ ] <!-- unlimit-branch=renovate/npm-9.x -->chore(deps): update npm to v9 - [ ] <!-- unlimit-branch=renovate/major-typescript-eslint-monorepo -->chore(deps): update typescript-eslint monorepo to v6 (major) (`@typescript-eslint/eslint-plugin`, `@typescript-eslint/parser`) - [ ] <!-- unlimit-branch=renovate/major-fontsource-monorepo -->fix(deps): update dependency @fontsource/roboto to v5 - [ ] <!-- unlimit-branch=renovate/google-cloud-functions-framework-3.x -->fix(deps): update dependency @google-cloud/functions-framework to v3 - [ ] <!-- unlimit-branch=renovate/testing-library-jest-dom-6.x -->fix(deps): update dependency @testing-library/jest-dom to v6 - [ ] <!-- unlimit-branch=renovate/testing-library-react-14.x -->fix(deps): update dependency @testing-library/react to v14 - [ ] <!-- unlimit-branch=renovate/testing-library-user-event-14.x -->fix(deps): update dependency @testing-library/user-event to v14 - [ ] <!-- unlimit-branch=renovate/axios-1.x -->fix(deps): update dependency axios to v1 - [ ] <!-- unlimit-branch=renovate/react-pdf-7.x -->fix(deps): update dependency react-pdf to v7 - [ ] <!-- unlimit-branch=renovate/web-vitals-3.x -->fix(deps): update dependency web-vitals to v3 - [ ] <!-- create-all-rate-limited-prs -->🔐 **Create all rate-limited PRs at once** 🔐 ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/eslint-config-prettier-8.x-lockfile -->[chore(deps): update dependency eslint-config-prettier to v8.10.0](../pull/505) - [ ] <!-- rebase-branch=renovate/eslint-8.x-lockfile -->[fix(deps): update dependency eslint to v8.47.0](../pull/507) ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/google-4.x-lockfile -->[chore(deps): update terraform google to v4.78.0](../pull/506) - [ ] <!-- rebase-branch=renovate/flowbite-svelte-0.x -->[fix(deps): update dependency flowbite-svelte to ^0.43.0](../pull/508) - [ ] <!-- rebase-branch=renovate/actions-checkout-3.x -->[chore(deps): update actions/checkout action to v3](../pull/509) - [ ] <!-- rebase-branch=renovate/coverage-7.x -->[chore(deps): update dependency coverage to v7.3.0](../pull/510) - [ ] <!-- rebase-branch=renovate/eslint-config-prettier-9.x -->[chore(deps): update dependency eslint-config-prettier to v9](../pull/511) - [ ] <!-- rebase-branch=renovate/eslint-plugin-prettier-5.x -->[chore(deps): update dependency eslint-plugin-prettier to v5](../pull/512) - [ ] <!-- rebase-branch=renovate/gunicorn-21.x -->[chore(deps): update dependency gunicorn to v21](../pull/514) - [ ] <!-- rebase-branch=renovate/mock-5.x -->[chore(deps): update dependency mock to v5.1.0](../pull/515) - [ ] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once** ## Detected dependencies <details><summary>dockerfile</summary> <blockquote> <details><summary>dialogflow-cx/vpc-sc-auth-server/server/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/Dockerfile.debug</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/agent/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/proxy-server-src/Dockerfile</summary> - `python 3.11.4-slim` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/Dockerfile</summary> </details> </blockquote> </details> <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/divebooker-agent.yaml</summary> - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` </details> <details><summary>.github/workflows/lint-markdown.yaml</summary> - `actions/checkout v2` - `DavidAnson/markdownlint-cli2-action v11` </details> <details><summary>.github/workflows/lint-terraform.yaml</summary> - `actions/checkout v3` - `terraform-linters/setup-tflint v3` </details> <details><summary>.github/workflows/lint-yaml.yaml</summary> - `actions/checkout v2` - `ibiqlik/action-yamllint v3` </details> <details><summary>.github/workflows/scorecards.yml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `ossf/scorecard-action v2.2.0@08b4669551908b1024bb425080c797723083c031` - `actions/upload-artifact v3.1.2@0b7f8abb1508181956e8e162db84b466c27e18ce` - `github/codeql-action v2.21.4@a09933a12a80f87b87005513f0abb1494c27a716` </details> <details><summary>.github/workflows/shirt-order-agent-terraform.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-auth-server.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-backend.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-frontend.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `actions/setup-node v3` </details> <details><summary>.github/workflows/vpc-sc-demo-reverse-proxy.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-terraform.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-webhook.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> </blockquote> </details> <details><summary>npm</summary> <blockquote> <details><summary>dialogflow-cx-nodejs/package.json</summary> - `@google-cloud/dialogflow-cx 4.0.0` - `axios 1.4.0` - `uuid 9.0.0` - `chai 4.3.7` - `eslint-plugin-json 3.1.0` - `mocha 10.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/frontend/package.json</summary> - `@emotion/react ^11.10.4` - `@emotion/styled ^11.10.4` - `@fontsource/roboto ^4.5.8` - `@mui/icons-material ^5.10.2` - `@mui/material ^5.10.3` - `@testing-library/jest-dom ^5.16.5` - `@testing-library/react ^13.3.0` - `@testing-library/user-event ^13.5.0` - `axios ^1.0.0` - `eslint ^8.30.0` - `gapi-script ^1.2.0` - `http-proxy-middleware ^2.0.6` - `react ^18.2.0` - `react-dom ^18.2.0` - `react-pdf ^5.7.2` - `@tanstack/react-query ^4.0.5` - `react-router-dom ^6.3.0` - `react-syntax-highlighter ^15.5.0` - `typeface-roboto ^1.1.13` - `web-vitals ^2.1.4` - `eslint-config-prettier ^8.5.0` - `eslint-plugin-json ^3.1.0` - `eslint-plugin-node ^11.1.0` - `eslint-plugin-prettier ^4.2.1` - `eslint-plugin-react ^7.31.11` - `prettier ^2.8.1` - `react-scripts ^5.0.1` - `stylelint ^15.10.1` - `stylelint-config-standard ^29.0.0` - `node 18.17.1` - `npm 8.19.4` </details> <details><summary>generative-chatbots/package.json</summary> - `@popperjs/core ^2.11.8` - `flowbite ^1.6.5` - `flowbite-svelte ^0.38.1` - `tailwind-merge ^1.13.1` - `@sveltejs/adapter-static ^2.0.2` - `@sveltejs/kit ^1.5.0` - `@typescript-eslint/eslint-plugin ^5.45.0` - `@typescript-eslint/parser ^5.45.0` - `autoprefixer ^10.4.14` - `eslint ^8.28.0` - `eslint-config-prettier ^8.5.0` - `eslint-plugin-svelte ^2.26.0` - `postcss ^8.4.23` - `postcss-load-config ^4.0.1` - `prettier ^3.0.2` - `prettier-plugin-svelte ^3.0.3` - `prettier-plugin-tailwindcss ^0.5.3` - `svelte ^3.54.0` - `svelte-check ^3.0.1` - `svelte-preprocess ^5.0.3` - `tailwindcss ^3.3.1` - `tslib ^2.4.1` - `typescript ^5.0.0` - `vite ^4.3.9` </details> <details><summary>goodbye-graffiti-agent/maps-function/package.json</summary> - `@google-cloud/functions-framework ^2.0.0` - `axios ^0.24.0` </details> <details><summary>package.json</summary> - `@google-cloud/dialogflow-cx 4.0.0` - `uuid 9.0.0` - `@typescript-eslint/eslint-plugin ^5.62.0` - `eslint ^8.47.0` - `eslint-config-prettier ^9.0.0` - `eslint-config-standard-with-typescript ^37.0.0` - `eslint-plugin-import ^2.28.0` - `eslint-plugin-json 3.1.0` - `eslint-plugin-n ^16.0.1` - `eslint-plugin-node ^11.1.0` - `eslint-plugin-prettier ^5.0.0` - `eslint-plugin-promise ^6.1.1` - `eslint-plugin-react ^7.33.1` - `eslint-plugin-react-hooks ^4.6.0` - `typescript ^5.1.6` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>dialogflow-cx/requirements-test.txt</summary> - `pytest ==7.4.0` - `pytest-rerunfailures ==12.0` </details> <details><summary>dialogflow-cx/requirements.txt</summary> - `google-cloud-dialogflow-cx ==1.26.0` - `google-auth ==2.22.0` - `grpcio ==1.57.0` - `mock ==5.0.2` </details> <details><summary>dialogflow-cx/vpc-sc-auth-server/server/requirements-test.txt</summary> - `pytest ==7.4.0` - `pytest-mypy ==0.10.3` - `mock ==5.0.2` - `invoke ==2.2.0` - `pyyaml ==6.0.1` - `coverage ==7.2.7` - `pytest-cov ==4.1.0` </details> <details><summary>dialogflow-cx/vpc-sc-auth-server/server/requirements.txt</summary> - `Flask ==2.3.2` - `werkzeug ==2.3.7` - `types-Flask ==1.1.6` - `Flask-Logging ==0.1.3` - `gunicorn ==20.1.0` - `requests ==2.31.0` - `google-auth ==2.22.0` - `google-cloud-secret-manager ==2.16.3` - `google-cloud-storage ==2.10.0` - `google-crc32c ==1.5.0` - `pycryptodome ==3.18.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/backend/requirements-test.txt</summary> - `pytest ==7.4.0` - `mock ==4.0.3` - `types-jsonschema ==4.17.0.10` </details> <details><summary>dialogflow-cx/vpc-sc-demo/backend/requirements.txt</summary> - `Flask ==2.3.2` - `Turbo-Flask ==0.8.4` - `gunicorn ==20.1.0` - `invoke ==2.2.0` - `requests ==2.31.0` - `google-auth ==2.22.0` - `google-crc32c ==1.5.0` - `pycryptodome ==3.18.0` - `google-cloud-storage ==2.10.0` - `invoke ==2.2.0` - `werkzeug ==2.3.7` - `google-cloud-bigquery ==3.11.4` - `jsonschema ==4.19.0` - `types-jsonschema ==4.17.0.10` - `google-cloud-bigquery ==3.11.4` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/proxy-server-src/requirements-test.txt</summary> - `pytest ==7.4.0` - `Flask ==2.3.2` - `werkzeug ==2.3.7` - `types-Flask ==1.1.6` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/proxy-server-src/requirements.txt</summary> - `Flask ==2.3.2` - `gunicorn ==20.1.0` - `requests ==2.31.0` - `structlog ==23.1.0` - `google-auth ==2.22.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/webhook/telecom-webhook-src/requirements-test.txt</summary> - `pytest ==7.4.0` - `Flask ==2.3.2` - `werkzeug ==2.3.7` - `types-Flask ==1.1.6` - `coverage ==6.5.0` - `pytest-cov ==4.1.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/requirements-test.txt</summary> - `pytest ==7.4.0` - `networkx ==2.8.8` - `pygraphviz ==1.11` </details> </blockquote> </details> <details><summary>terraform</summary> <blockquote> <details><summary>dialogflow-cx/shirt-order-agent/provider.tf</summary> - `google >= 4.40.0` - `null >= 3.2.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/terraform/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/agent/main.tf</summary> - `google >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `google-beta >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/service_directory/main.tf</summary> - `google >= 4.77.0` - `google-beta >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/vpc_network/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/webhook/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/service-directory/main.tf</summary> - `google >= 4.77.0` - `google-beta >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/service-perimeter/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/services/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/vpc-network/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `google-beta >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/webhook-agent/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>oidc/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` - `terraform-google-modules/github-actions-runners/google 3.1.1` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Rate-Limited These updates are currently rate-limited. Click on a checkbox below to force their creation now. - [ ] <!-- unlimit-branch=renovate/prettier-2.x-lockfile -->chore(deps): update dependency prettier to v2.8.8 - [ ] <!-- unlimit-branch=renovate/stylelint-15.x-lockfile -->chore(deps): update dependency stylelint to v15.10.2 - [ ] <!-- unlimit-branch=renovate/eslint-plugin-react-7.x-lockfile -->chore(deps): update dependency eslint-plugin-react to v7.33.2 - [ ] <!-- unlimit-branch=renovate/tanstack-query-monorepo -->fix(deps): update dependency @tanstack/react-query to v4.32.6 - [ ] <!-- unlimit-branch=renovate/testing-library-jest-dom-5.x-lockfile -->fix(deps): update dependency @testing-library/jest-dom to v5.17.0 - [ ] <!-- unlimit-branch=renovate/axios-0.x -->fix(deps): update dependency axios to ^0.27.0 - [ ] <!-- unlimit-branch=renovate/react-router-monorepo -->fix(deps): update dependency react-router-dom to v6.15.0 - [ ] <!-- unlimit-branch=renovate/emotion-monorepo -->fix(deps): update emotion monorepo to v11.11.1 (`@emotion/react`, `@emotion/styled`) - [ ] <!-- unlimit-branch=renovate/material-ui-monorepo -->fix(deps): update material-ui monorepo (`@mui/icons-material`, `@mui/material`) - [ ] <!-- unlimit-branch=renovate/networkx-3.x -->chore(deps): update dependency networkx to v3 - [ ] <!-- unlimit-branch=renovate/prettier-3.x -->chore(deps): update dependency prettier to v3 - [ ] <!-- unlimit-branch=renovate/stylelint-config-standard-34.x -->chore(deps): update dependency stylelint-config-standard to v34 - [ ] <!-- unlimit-branch=renovate/svelte-4.x -->chore(deps): update dependency svelte to v4 - [ ] <!-- unlimit-branch=renovate/npm-9.x -->chore(deps): update npm to v9 - [ ] <!-- unlimit-branch=renovate/major-typescript-eslint-monorepo -->chore(deps): update typescript-eslint monorepo to v6 (major) (`@typescript-eslint/eslint-plugin`, `@typescript-eslint/parser`) - [ ] <!-- unlimit-branch=renovate/major-fontsource-monorepo -->fix(deps): update dependency @fontsource/roboto to v5 - [ ] <!-- unlimit-branch=renovate/google-cloud-functions-framework-3.x -->fix(deps): update dependency @google-cloud/functions-framework to v3 - [ ] <!-- unlimit-branch=renovate/testing-library-jest-dom-6.x -->fix(deps): update dependency @testing-library/jest-dom to v6 - [ ] <!-- unlimit-branch=renovate/testing-library-react-14.x -->fix(deps): update dependency @testing-library/react to v14 - [ ] <!-- unlimit-branch=renovate/testing-library-user-event-14.x -->fix(deps): update dependency @testing-library/user-event to v14 - [ ] <!-- unlimit-branch=renovate/axios-1.x -->fix(deps): update dependency axios to v1 - [ ] <!-- unlimit-branch=renovate/react-pdf-7.x -->fix(deps): update dependency react-pdf to v7 - [ ] <!-- unlimit-branch=renovate/web-vitals-3.x -->fix(deps): update dependency web-vitals to v3 - [ ] <!-- create-all-rate-limited-prs -->🔐 **Create all rate-limited PRs at once** 🔐 ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/eslint-config-prettier-8.x-lockfile -->[chore(deps): update dependency eslint-config-prettier to v8.10.0](../pull/505) - [ ] <!-- rebase-branch=renovate/eslint-8.x-lockfile -->[fix(deps): update dependency eslint to v8.47.0](../pull/507) ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/google-4.x-lockfile -->[chore(deps): update terraform google to v4.78.0](../pull/506) - [ ] <!-- rebase-branch=renovate/flowbite-svelte-0.x -->[fix(deps): update dependency flowbite-svelte to ^0.43.0](../pull/508) - [ ] <!-- rebase-branch=renovate/actions-checkout-3.x -->[chore(deps): update actions/checkout action to v3](../pull/509) - [ ] <!-- rebase-branch=renovate/coverage-7.x -->[chore(deps): update dependency coverage to v7.3.0](../pull/510) - [ ] <!-- rebase-branch=renovate/eslint-config-prettier-9.x -->[chore(deps): update dependency eslint-config-prettier to v9](../pull/511) - [ ] <!-- rebase-branch=renovate/eslint-plugin-prettier-5.x -->[chore(deps): update dependency eslint-plugin-prettier to v5](../pull/512) - [ ] <!-- rebase-branch=renovate/gunicorn-21.x -->[chore(deps): update dependency gunicorn to v21](../pull/514) - [ ] <!-- rebase-branch=renovate/mock-5.x -->[chore(deps): update dependency mock to v5.1.0](../pull/515) - [ ] <!-- rebase-all-open-prs -->**Click on this checkbox to rebase all open PRs at once** ## Detected dependencies <details><summary>dockerfile</summary> <blockquote> <details><summary>dialogflow-cx/vpc-sc-auth-server/server/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/Dockerfile.debug</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/agent/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/Dockerfile</summary> </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/proxy-server-src/Dockerfile</summary> - `python 3.11.4-slim` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/Dockerfile</summary> </details> </blockquote> </details> <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/divebooker-agent.yaml</summary> - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` </details> <details><summary>.github/workflows/lint-markdown.yaml</summary> - `actions/checkout v2` - `DavidAnson/markdownlint-cli2-action v11` </details> <details><summary>.github/workflows/lint-terraform.yaml</summary> - `actions/checkout v3` - `terraform-linters/setup-tflint v3` </details> <details><summary>.github/workflows/lint-yaml.yaml</summary> - `actions/checkout v2` - `ibiqlik/action-yamllint v3` </details> <details><summary>.github/workflows/scorecards.yml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `ossf/scorecard-action v2.2.0@08b4669551908b1024bb425080c797723083c031` - `actions/upload-artifact v3.1.2@0b7f8abb1508181956e8e162db84b466c27e18ce` - `github/codeql-action v2.21.4@a09933a12a80f87b87005513f0abb1494c27a716` </details> <details><summary>.github/workflows/shirt-order-agent-terraform.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-auth-server.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-backend.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-frontend.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `actions/setup-node v3` </details> <details><summary>.github/workflows/vpc-sc-demo-reverse-proxy.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-terraform.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> <details><summary>.github/workflows/vpc-sc-demo-webhook.yaml</summary> - `step-security/harden-runner v2.5.1@8ca2b8b2ece13480cda6dacd3511b49857a23c09` - `actions/checkout v3.5.3@c85c95e3d7251135ab7dc9ce3241c5835cc595a9` - `codecov/codecov-action v3` </details> </blockquote> </details> <details><summary>npm</summary> <blockquote> <details><summary>dialogflow-cx-nodejs/package.json</summary> - `@google-cloud/dialogflow-cx 4.0.0` - `axios 1.4.0` - `uuid 9.0.0` - `chai 4.3.7` - `eslint-plugin-json 3.1.0` - `mocha 10.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/frontend/package.json</summary> - `@emotion/react ^11.10.4` - `@emotion/styled ^11.10.4` - `@fontsource/roboto ^4.5.8` - `@mui/icons-material ^5.10.2` - `@mui/material ^5.10.3` - `@testing-library/jest-dom ^5.16.5` - `@testing-library/react ^13.3.0` - `@testing-library/user-event ^13.5.0` - `axios ^1.0.0` - `eslint ^8.30.0` - `gapi-script ^1.2.0` - `http-proxy-middleware ^2.0.6` - `react ^18.2.0` - `react-dom ^18.2.0` - `react-pdf ^5.7.2` - `@tanstack/react-query ^4.0.5` - `react-router-dom ^6.3.0` - `react-syntax-highlighter ^15.5.0` - `typeface-roboto ^1.1.13` - `web-vitals ^2.1.4` - `eslint-config-prettier ^8.5.0` - `eslint-plugin-json ^3.1.0` - `eslint-plugin-node ^11.1.0` - `eslint-plugin-prettier ^4.2.1` - `eslint-plugin-react ^7.31.11` - `prettier ^2.8.1` - `react-scripts ^5.0.1` - `stylelint ^15.10.1` - `stylelint-config-standard ^29.0.0` - `node 18.17.1` - `npm 8.19.4` </details> <details><summary>generative-chatbots/package.json</summary> - `@popperjs/core ^2.11.8` - `flowbite ^1.6.5` - `flowbite-svelte ^0.38.1` - `tailwind-merge ^1.13.1` - `@sveltejs/adapter-static ^2.0.2` - `@sveltejs/kit ^1.5.0` - `@typescript-eslint/eslint-plugin ^5.45.0` - `@typescript-eslint/parser ^5.45.0` - `autoprefixer ^10.4.14` - `eslint ^8.28.0` - `eslint-config-prettier ^8.5.0` - `eslint-plugin-svelte ^2.26.0` - `postcss ^8.4.23` - `postcss-load-config ^4.0.1` - `prettier ^3.0.2` - `prettier-plugin-svelte ^3.0.3` - `prettier-plugin-tailwindcss ^0.5.3` - `svelte ^3.54.0` - `svelte-check ^3.0.1` - `svelte-preprocess ^5.0.3` - `tailwindcss ^3.3.1` - `tslib ^2.4.1` - `typescript ^5.0.0` - `vite ^4.3.9` </details> <details><summary>goodbye-graffiti-agent/maps-function/package.json</summary> - `@google-cloud/functions-framework ^2.0.0` - `axios ^0.24.0` </details> <details><summary>package.json</summary> - `@google-cloud/dialogflow-cx 4.0.0` - `uuid 9.0.0` - `@typescript-eslint/eslint-plugin ^5.62.0` - `eslint ^8.47.0` - `eslint-config-prettier ^9.0.0` - `eslint-config-standard-with-typescript ^37.0.0` - `eslint-plugin-import ^2.28.0` - `eslint-plugin-json 3.1.0` - `eslint-plugin-n ^16.0.1` - `eslint-plugin-node ^11.1.0` - `eslint-plugin-prettier ^5.0.0` - `eslint-plugin-promise ^6.1.1` - `eslint-plugin-react ^7.33.1` - `eslint-plugin-react-hooks ^4.6.0` - `typescript ^5.1.6` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>dialogflow-cx/requirements-test.txt</summary> - `pytest ==7.4.0` - `pytest-rerunfailures ==12.0` </details> <details><summary>dialogflow-cx/requirements.txt</summary> - `google-cloud-dialogflow-cx ==1.26.0` - `google-auth ==2.22.0` - `grpcio ==1.57.0` - `mock ==5.0.2` </details> <details><summary>dialogflow-cx/vpc-sc-auth-server/server/requirements-test.txt</summary> - `pytest ==7.4.0` - `pytest-mypy ==0.10.3` - `mock ==5.0.2` - `invoke ==2.2.0` - `pyyaml ==6.0.1` - `coverage ==7.2.7` - `pytest-cov ==4.1.0` </details> <details><summary>dialogflow-cx/vpc-sc-auth-server/server/requirements.txt</summary> - `Flask ==2.3.2` - `werkzeug ==2.3.7` - `types-Flask ==1.1.6` - `Flask-Logging ==0.1.3` - `gunicorn ==20.1.0` - `requests ==2.31.0` - `google-auth ==2.22.0` - `google-cloud-secret-manager ==2.16.3` - `google-cloud-storage ==2.10.0` - `google-crc32c ==1.5.0` - `pycryptodome ==3.18.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/backend/requirements-test.txt</summary> - `pytest ==7.4.0` - `mock ==4.0.3` - `types-jsonschema ==4.17.0.10` </details> <details><summary>dialogflow-cx/vpc-sc-demo/backend/requirements.txt</summary> - `Flask ==2.3.2` - `Turbo-Flask ==0.8.4` - `gunicorn ==20.1.0` - `invoke ==2.2.0` - `requests ==2.31.0` - `google-auth ==2.22.0` - `google-crc32c ==1.5.0` - `pycryptodome ==3.18.0` - `google-cloud-storage ==2.10.0` - `invoke ==2.2.0` - `werkzeug ==2.3.7` - `google-cloud-bigquery ==3.11.4` - `jsonschema ==4.19.0` - `types-jsonschema ==4.17.0.10` - `google-cloud-bigquery ==3.11.4` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/proxy-server-src/requirements-test.txt</summary> - `pytest ==7.4.0` - `Flask ==2.3.2` - `werkzeug ==2.3.7` - `types-Flask ==1.1.6` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/proxy-server-src/requirements.txt</summary> - `Flask ==2.3.2` - `gunicorn ==20.1.0` - `requests ==2.31.0` - `structlog ==23.1.0` - `google-auth ==2.22.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/webhook/telecom-webhook-src/requirements-test.txt</summary> - `pytest ==7.4.0` - `Flask ==2.3.2` - `werkzeug ==2.3.7` - `types-Flask ==1.1.6` - `coverage ==6.5.0` - `pytest-cov ==4.1.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/requirements-test.txt</summary> - `pytest ==7.4.0` - `networkx ==2.8.8` - `pygraphviz ==1.11` </details> </blockquote> </details> <details><summary>terraform</summary> <blockquote> <details><summary>dialogflow-cx/shirt-order-agent/provider.tf</summary> - `google >= 4.40.0` - `null >= 3.2.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/terraform/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/agent/main.tf</summary> - `google >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/reverse_proxy_server/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `google-beta >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/service_directory/main.tf</summary> - `google >= 4.77.0` - `google-beta >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/vpc_network/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/components/webhook/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/service-directory/main.tf</summary> - `google >= 4.77.0` - `google-beta >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/service-perimeter/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/services/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/vpc-network/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `google-beta >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>dialogflow-cx/vpc-sc-demo/deploy/terraform/webhook-agent/main.tf</summary> - `archive >= 2.4.0` - `google >= 4.77.0` - `time >= 0.9.1` - `hashicorp/terraform >= 1.2.0` </details> <details><summary>oidc/main.tf</summary> - `google >= 4.77.0` - `hashicorp/terraform >= 1.2.0` - `terraform-google-modules/github-actions-runners/google 3.1.1` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more rate limited these updates are currently rate limited click on a checkbox below to force their creation now chore deps update dependency prettier to chore deps update dependency stylelint to chore deps update dependency eslint plugin react to fix deps update dependency tanstack react query to fix deps update dependency testing library jest dom to fix deps update dependency axios to fix deps update dependency react router dom to fix deps update emotion monorepo to emotion react emotion styled fix deps update material ui monorepo mui icons material mui material chore deps update dependency networkx to chore deps update dependency prettier to chore deps update dependency stylelint config standard to chore deps update dependency svelte to chore deps update npm to chore deps update typescript eslint monorepo to major typescript eslint eslint plugin typescript eslint parser fix deps update dependency fontsource roboto to fix deps update dependency google cloud functions framework to fix deps update dependency testing library jest dom to fix deps update dependency testing library react to fix deps update dependency testing library user event to fix deps update dependency axios to fix deps update dependency react pdf to fix deps update dependency web vitals to 🔐 create all rate limited prs at once 🔐 edited blocked these updates have been manually edited so renovate will no longer make changes to discard all commits and start over click on a checkbox pull pull open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull pull pull pull pull pull pull click on this checkbox to rebase all open prs at once detected dependencies dockerfile dialogflow cx vpc sc auth server server dockerfile dialogflow cx vpc sc demo dockerfile dialogflow cx vpc sc demo dockerfile debug dialogflow cx vpc sc demo components agent dockerfile dialogflow cx vpc sc demo components reverse proxy server dockerfile dialogflow cx vpc sc demo components reverse proxy server proxy server src dockerfile python slim dialogflow cx vpc sc demo deploy terraform dockerfile github actions github workflows divebooker agent yaml actions checkout github workflows lint markdown yaml actions checkout davidanson markdownlint action github workflows lint terraform yaml actions checkout terraform linters setup tflint github workflows lint yaml yaml actions checkout ibiqlik action yamllint github workflows scorecards yml step security harden runner actions checkout ossf scorecard action actions upload artifact github codeql action github workflows shirt order agent terraform yaml step security harden runner actions checkout codecov codecov action github workflows vpc sc auth server yaml step security harden runner actions checkout codecov codecov action github workflows vpc sc demo backend yaml step security harden runner actions checkout codecov codecov action github workflows vpc sc demo frontend yaml step security harden runner actions checkout actions setup node github workflows vpc sc demo reverse proxy yaml step security harden runner actions checkout codecov codecov action github workflows vpc sc demo terraform yaml step security harden runner actions checkout codecov codecov action github workflows vpc sc demo webhook yaml step security harden runner actions checkout codecov codecov action npm dialogflow cx nodejs package json google cloud dialogflow cx axios uuid chai eslint plugin json mocha dialogflow cx vpc sc demo frontend package json emotion react emotion styled fontsource roboto mui icons material mui material testing library jest dom testing library react testing library user event axios eslint gapi script http proxy middleware react react dom react pdf tanstack react query react router dom react syntax highlighter typeface roboto web vitals eslint config prettier eslint plugin json eslint plugin node eslint plugin prettier eslint plugin react prettier react scripts stylelint stylelint config standard node npm generative chatbots package json popperjs core flowbite flowbite svelte tailwind merge sveltejs adapter static sveltejs kit typescript eslint eslint plugin typescript eslint parser autoprefixer eslint eslint config prettier eslint plugin svelte postcss postcss load config prettier prettier plugin svelte prettier plugin tailwindcss svelte svelte check svelte preprocess tailwindcss tslib typescript vite goodbye graffiti agent maps function package json google cloud functions framework axios package json google cloud dialogflow cx uuid typescript eslint eslint plugin eslint eslint config prettier eslint config standard with typescript eslint plugin import eslint plugin json eslint plugin n eslint plugin node eslint plugin prettier eslint plugin promise eslint plugin react eslint plugin react hooks typescript pip requirements dialogflow cx requirements test txt pytest pytest rerunfailures dialogflow cx requirements txt google cloud dialogflow cx google auth grpcio mock dialogflow cx vpc sc auth server server requirements test txt pytest pytest mypy mock invoke pyyaml coverage pytest cov dialogflow cx vpc sc auth server server requirements txt flask werkzeug types flask flask logging gunicorn requests google auth google cloud secret manager google cloud storage google pycryptodome dialogflow cx vpc sc demo backend requirements test txt pytest mock types jsonschema dialogflow cx vpc sc demo backend requirements txt flask turbo flask gunicorn invoke requests google auth google pycryptodome google cloud storage invoke werkzeug google cloud bigquery jsonschema types jsonschema google cloud bigquery dialogflow cx vpc sc demo components reverse proxy server proxy server src requirements test txt pytest flask werkzeug types flask dialogflow cx vpc sc demo components reverse proxy server proxy server src requirements txt flask gunicorn requests structlog google auth dialogflow cx vpc sc demo components webhook telecom webhook src requirements test txt pytest flask werkzeug types flask coverage pytest cov dialogflow cx vpc sc demo deploy terraform requirements test txt pytest networkx pygraphviz terraform dialogflow cx shirt order agent provider tf google null hashicorp terraform dialogflow cx terraform main tf archive google hashicorp terraform dialogflow cx vpc sc demo components agent main tf google time hashicorp terraform dialogflow cx vpc sc demo components reverse proxy server main tf archive google google beta time hashicorp terraform dialogflow cx vpc sc demo components service directory main tf google google beta hashicorp terraform dialogflow cx vpc sc demo components vpc network main tf google hashicorp terraform dialogflow cx vpc sc demo components webhook main tf archive google time hashicorp terraform dialogflow cx vpc sc demo deploy terraform main tf google hashicorp terraform dialogflow cx vpc sc demo deploy terraform service directory main tf google google beta hashicorp terraform dialogflow cx vpc sc demo deploy terraform service perimeter main tf google hashicorp terraform dialogflow cx vpc sc demo deploy terraform services main tf google hashicorp terraform dialogflow cx vpc sc demo deploy terraform vpc network main tf archive google google beta time hashicorp terraform dialogflow cx vpc sc demo deploy terraform webhook agent main tf archive google time hashicorp terraform oidc main tf google hashicorp terraform terraform google modules github actions runners google check this box to trigger a request for renovate to run again on this repository
1
587,157
17,605,889,436
IssuesEvent
2021-08-17 17:00:48
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Blazor client developers can use Parallel.For in their applications to share more code with server
arch-wasm area-System.Threading.Tasks User Story Priority:2
<!--This is just a template - feel free to delete any and all of it and replace as appropriate.--> ### Description <!-- * Please share a clear and concise description of the problem. * Include minimal steps to reproduce the problem if possible. E.g.: the smallest possible code snippet; or a small repo to clone, with steps to run it. * What behavior are you seeing, and what behavior would you expect? --> Hi, So the Roslyn compiler is kind of working in blazor wasm with the latest version (3.8.0-4.final) which is great! Appears the hashing issues are fixed. However, now I find there is a strange bug whereby I receive a multi threading exception under the following scenario: - Dotnet run - Load web page in browser - Run compile (works fine) - Refresh page in browser - Run compile (exception) Strangely it works if I close the original tab and open a new one. Thanks ### Configuration <!-- * Which version of .NET is the code running on? * What OS and version, and what distro if applicable? * What is the architecture (x64, x86, ARM, ARM64)? * Do you know whether it is specific to that configuration? * If you're using Blazor, which web browser(s) do you see this issue in? --> - .NET 5.0.100-rc.2.20479.15 - Blazor wasm - Microsoft.CodeAnalysis 3.8.0-4.final - Chrome ### Regression? <!-- * Did this work in a previous build or release of .NET Core, or from .NET Framework? If you can try a previous release or build to find out, that can help us narrow down the problem. If you don't know, that's OK. --> Last time Roslyn was working in the browser was with Blazor Webassembly 3.2.1 and Microsoft.CodeAnalysis 3.6.0 ### Other information <!-- * Please include any relevant stack traces or error messages. If possible please include text as text rather than images (so it shows up in searches). * If you have an idea where the problem might lie, let us know that here. Please include any pointers to code, relevant changes, or related issues you know of. * Do you know of any workarounds? --> ``` System.PlatformNotSupportedException: Cannot wait on monitors on this runtime. at System.Threading.Monitor.ObjWait(Boolean exitContext, Int32 millisecondsTimeout, Object obj) at System.Threading.Monitor.Wait(Object obj, Int32 millisecondsTimeout, Boolean exitContext) at System.Threading.Monitor.Wait(Object obj, Int32 millisecondsTimeout) at System.Threading.ManualResetEventSlim.Wait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.SpinThenBlockingWait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.InternalWaitCore(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.InternalWait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.Wait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.Wait() at System.Threading.Tasks.TaskReplicator.Replica.Wait() at System.Threading.Tasks.TaskReplicator.Run[RangeWorker](ReplicatableUserAction`1 action, ParallelOptions options, Boolean stopOnFirstFailure) at System.Threading.Tasks.Parallel.ForWorker[Object](Int32 fromInclusive, Int32 toExclusive, ParallelOptions parallelOptions, Action`1 body, Action`2 bodyWithState, Func`4 bodyWithLocal, Func`1 localInit, Action`1 localFinally) at System.Threading.Tasks.Parallel.For(Int32 fromInclusive, Int32 toExclusive, ParallelOptions parallelOptions, Action`1 body) at Microsoft.CodeAnalysis.CSharp.CSharpCompilation.GetDiagnostics(CompilationStage stage, Boolean includeEarlierStages, DiagnosticBag diagnostics, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.CSharp.CSharpCompilation.GetDiagnostics(CompilationStage stage, Boolean includeEarlierStages, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.CSharp.CSharpCompilation.CompileMethods(CommonPEModuleBuilder moduleBuilder, Boolean emittingPdb, Boolean emitMetadataOnly, Boolean emitTestCoverageData, DiagnosticBag diagnostics, Predicate`1 filterOpt, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.Compilation.Emit(Stream peStream, Stream metadataPEStream, Stream pdbStream, Stream xmlDocumentationStream, Stream win32Resources, IEnumerable`1 manifestResources, EmitOptions options, IMethodSymbol debugEntryPoint, Stream sourceLinkStream, IEnumerable`1 embeddedTexts, CompilationTestData testData, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.Compilation.Emit(Stream peStream, Stream pdbStream, Stream xmlDocumentationStream, Stream win32Resources, IEnumerable`1 manifestResources, EmitOptions options, IMethodSymbol debugEntryPoint, Stream sourceLinkStream, IEnumerable`1 embeddedTexts, Stream metadataPEStream, CancellationToken cancellationToken) ```
1.0
Blazor client developers can use Parallel.For in their applications to share more code with server - <!--This is just a template - feel free to delete any and all of it and replace as appropriate.--> ### Description <!-- * Please share a clear and concise description of the problem. * Include minimal steps to reproduce the problem if possible. E.g.: the smallest possible code snippet; or a small repo to clone, with steps to run it. * What behavior are you seeing, and what behavior would you expect? --> Hi, So the Roslyn compiler is kind of working in blazor wasm with the latest version (3.8.0-4.final) which is great! Appears the hashing issues are fixed. However, now I find there is a strange bug whereby I receive a multi threading exception under the following scenario: - Dotnet run - Load web page in browser - Run compile (works fine) - Refresh page in browser - Run compile (exception) Strangely it works if I close the original tab and open a new one. Thanks ### Configuration <!-- * Which version of .NET is the code running on? * What OS and version, and what distro if applicable? * What is the architecture (x64, x86, ARM, ARM64)? * Do you know whether it is specific to that configuration? * If you're using Blazor, which web browser(s) do you see this issue in? --> - .NET 5.0.100-rc.2.20479.15 - Blazor wasm - Microsoft.CodeAnalysis 3.8.0-4.final - Chrome ### Regression? <!-- * Did this work in a previous build or release of .NET Core, or from .NET Framework? If you can try a previous release or build to find out, that can help us narrow down the problem. If you don't know, that's OK. --> Last time Roslyn was working in the browser was with Blazor Webassembly 3.2.1 and Microsoft.CodeAnalysis 3.6.0 ### Other information <!-- * Please include any relevant stack traces or error messages. If possible please include text as text rather than images (so it shows up in searches). * If you have an idea where the problem might lie, let us know that here. Please include any pointers to code, relevant changes, or related issues you know of. * Do you know of any workarounds? --> ``` System.PlatformNotSupportedException: Cannot wait on monitors on this runtime. at System.Threading.Monitor.ObjWait(Boolean exitContext, Int32 millisecondsTimeout, Object obj) at System.Threading.Monitor.Wait(Object obj, Int32 millisecondsTimeout, Boolean exitContext) at System.Threading.Monitor.Wait(Object obj, Int32 millisecondsTimeout) at System.Threading.ManualResetEventSlim.Wait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.SpinThenBlockingWait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.InternalWaitCore(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.InternalWait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.Wait(Int32 millisecondsTimeout, CancellationToken cancellationToken) at System.Threading.Tasks.Task.Wait() at System.Threading.Tasks.TaskReplicator.Replica.Wait() at System.Threading.Tasks.TaskReplicator.Run[RangeWorker](ReplicatableUserAction`1 action, ParallelOptions options, Boolean stopOnFirstFailure) at System.Threading.Tasks.Parallel.ForWorker[Object](Int32 fromInclusive, Int32 toExclusive, ParallelOptions parallelOptions, Action`1 body, Action`2 bodyWithState, Func`4 bodyWithLocal, Func`1 localInit, Action`1 localFinally) at System.Threading.Tasks.Parallel.For(Int32 fromInclusive, Int32 toExclusive, ParallelOptions parallelOptions, Action`1 body) at Microsoft.CodeAnalysis.CSharp.CSharpCompilation.GetDiagnostics(CompilationStage stage, Boolean includeEarlierStages, DiagnosticBag diagnostics, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.CSharp.CSharpCompilation.GetDiagnostics(CompilationStage stage, Boolean includeEarlierStages, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.CSharp.CSharpCompilation.CompileMethods(CommonPEModuleBuilder moduleBuilder, Boolean emittingPdb, Boolean emitMetadataOnly, Boolean emitTestCoverageData, DiagnosticBag diagnostics, Predicate`1 filterOpt, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.Compilation.Emit(Stream peStream, Stream metadataPEStream, Stream pdbStream, Stream xmlDocumentationStream, Stream win32Resources, IEnumerable`1 manifestResources, EmitOptions options, IMethodSymbol debugEntryPoint, Stream sourceLinkStream, IEnumerable`1 embeddedTexts, CompilationTestData testData, CancellationToken cancellationToken) at Microsoft.CodeAnalysis.Compilation.Emit(Stream peStream, Stream pdbStream, Stream xmlDocumentationStream, Stream win32Resources, IEnumerable`1 manifestResources, EmitOptions options, IMethodSymbol debugEntryPoint, Stream sourceLinkStream, IEnumerable`1 embeddedTexts, Stream metadataPEStream, CancellationToken cancellationToken) ```
non_process
blazor client developers can use parallel for in their applications to share more code with server description please share a clear and concise description of the problem include minimal steps to reproduce the problem if possible e g the smallest possible code snippet or a small repo to clone with steps to run it what behavior are you seeing and what behavior would you expect hi so the roslyn compiler is kind of working in blazor wasm with the latest version final which is great appears the hashing issues are fixed however now i find there is a strange bug whereby i receive a multi threading exception under the following scenario dotnet run load web page in browser run compile works fine refresh page in browser run compile exception strangely it works if i close the original tab and open a new one thanks configuration which version of net is the code running on what os and version and what distro if applicable what is the architecture arm do you know whether it is specific to that configuration if you re using blazor which web browser s do you see this issue in net rc blazor wasm microsoft codeanalysis final chrome regression did this work in a previous build or release of net core or from net framework if you can try a previous release or build to find out that can help us narrow down the problem if you don t know that s ok last time roslyn was working in the browser was with blazor webassembly and microsoft codeanalysis other information please include any relevant stack traces or error messages if possible please include text as text rather than images so it shows up in searches if you have an idea where the problem might lie let us know that here please include any pointers to code relevant changes or related issues you know of do you know of any workarounds system platformnotsupportedexception cannot wait on monitors on this runtime at system threading monitor objwait boolean exitcontext millisecondstimeout object obj at system threading monitor wait object obj millisecondstimeout boolean exitcontext at system threading monitor wait object obj millisecondstimeout at system threading manualreseteventslim wait millisecondstimeout cancellationtoken cancellationtoken at system threading tasks task spinthenblockingwait millisecondstimeout cancellationtoken cancellationtoken at system threading tasks task internalwaitcore millisecondstimeout cancellationtoken cancellationtoken at system threading tasks task internalwait millisecondstimeout cancellationtoken cancellationtoken at system threading tasks task wait millisecondstimeout cancellationtoken cancellationtoken at system threading tasks task wait at system threading tasks taskreplicator replica wait at system threading tasks taskreplicator run replicatableuseraction action paralleloptions options boolean stoponfirstfailure at system threading tasks parallel forworker frominclusive toexclusive paralleloptions paralleloptions action body action bodywithstate func bodywithlocal func localinit action localfinally at system threading tasks parallel for frominclusive toexclusive paralleloptions paralleloptions action body at microsoft codeanalysis csharp csharpcompilation getdiagnostics compilationstage stage boolean includeearlierstages diagnosticbag diagnostics cancellationtoken cancellationtoken at microsoft codeanalysis csharp csharpcompilation getdiagnostics compilationstage stage boolean includeearlierstages cancellationtoken cancellationtoken at microsoft codeanalysis csharp csharpcompilation compilemethods commonpemodulebuilder modulebuilder boolean emittingpdb boolean emitmetadataonly boolean emittestcoveragedata diagnosticbag diagnostics predicate filteropt cancellationtoken cancellationtoken at microsoft codeanalysis compilation emit stream pestream stream metadatapestream stream pdbstream stream xmldocumentationstream stream ienumerable manifestresources emitoptions options imethodsymbol debugentrypoint stream sourcelinkstream ienumerable embeddedtexts compilationtestdata testdata cancellationtoken cancellationtoken at microsoft codeanalysis compilation emit stream pestream stream pdbstream stream xmldocumentationstream stream ienumerable manifestresources emitoptions options imethodsymbol debugentrypoint stream sourcelinkstream ienumerable embeddedtexts stream metadatapestream cancellationtoken cancellationtoken
0
132,851
10,771,564,204
IssuesEvent
2019-11-02 08:42:51
Skycoin-cp/skycoin
https://api.github.com/repos/Skycoin-cp/skycoin
closed
Flaky test in gnet/pool
tests
Excerpt: ``` [2018-08-30T03:47:01Z] ERROR [gnet]: daemon.gnet.ConnectionPool.BroadcastMessage error: Connection pool is empty [2018-08-30T03:47:01Z] ERROR [gnet]: req.Func daemon.gnet.ConnectionPool.BroadcastMessage failed: Connection pool is empty --- FAIL: TestPoolBroadcastMessage (0.11s) Error Trace: pool_test.go:1075 Error: Should be true [2018-08-30T03:47:01Z] INFO [gnet]: Listening for connections on 127.0.0.1:50823... [2018-08-30T03:47:01Z] INFO [gnet]: Connection pool closed [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown called [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown closed pool.quit [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown waiting for strandDone [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown closing the listener [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown disconnecting all connections [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown waiting for done --- FAIL: TestPoolReceiveMessage (0.20s) Error Trace: pool_test.go:1096 asm_amd64.s:1333 Error: Received unexpected error: listen tcp 127.0.0.1:50823: bind: address already in use FAIL ``` Happening on travis intermittently
1.0
Flaky test in gnet/pool - Excerpt: ``` [2018-08-30T03:47:01Z] ERROR [gnet]: daemon.gnet.ConnectionPool.BroadcastMessage error: Connection pool is empty [2018-08-30T03:47:01Z] ERROR [gnet]: req.Func daemon.gnet.ConnectionPool.BroadcastMessage failed: Connection pool is empty --- FAIL: TestPoolBroadcastMessage (0.11s) Error Trace: pool_test.go:1075 Error: Should be true [2018-08-30T03:47:01Z] INFO [gnet]: Listening for connections on 127.0.0.1:50823... [2018-08-30T03:47:01Z] INFO [gnet]: Connection pool closed [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown called [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown closed pool.quit [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown waiting for strandDone [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown closing the listener [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown disconnecting all connections [2018-08-30T03:47:02Z] INFO [gnet]: ConnectionPool.Shutdown waiting for done --- FAIL: TestPoolReceiveMessage (0.20s) Error Trace: pool_test.go:1096 asm_amd64.s:1333 Error: Received unexpected error: listen tcp 127.0.0.1:50823: bind: address already in use FAIL ``` Happening on travis intermittently
non_process
flaky test in gnet pool excerpt error daemon gnet connectionpool broadcastmessage error connection pool is empty error req func daemon gnet connectionpool broadcastmessage failed connection pool is empty fail testpoolbroadcastmessage error trace pool test go error should be true info listening for connections on info connection pool closed info connectionpool shutdown called info connectionpool shutdown closed pool quit info connectionpool shutdown waiting for stranddone info connectionpool shutdown closing the listener info connectionpool shutdown disconnecting all connections info connectionpool shutdown waiting for done fail testpoolreceivemessage error trace pool test go asm s error received unexpected error listen tcp bind address already in use fail happening on travis intermittently
0
1,070
2,713,435,055
IssuesEvent
2015-04-09 19:15:45
PeerioTechnologies/peerio-client
https://api.github.com/repos/PeerioTechnologies/peerio-client
opened
Pending contact requests should be bumped to top of list
usability
Invited contacts are bumped to the top of the list. Pending contacts (awaiting user to accept) seem to be sorted as approved contacts are. It would make more sense to put pending contacts at the top of the list (ones that require user-response), and put invited contacts at the bottom (no user response needed).
True
Pending contact requests should be bumped to top of list - Invited contacts are bumped to the top of the list. Pending contacts (awaiting user to accept) seem to be sorted as approved contacts are. It would make more sense to put pending contacts at the top of the list (ones that require user-response), and put invited contacts at the bottom (no user response needed).
non_process
pending contact requests should be bumped to top of list invited contacts are bumped to the top of the list pending contacts awaiting user to accept seem to be sorted as approved contacts are it would make more sense to put pending contacts at the top of the list ones that require user response and put invited contacts at the bottom no user response needed
0
19,600
25,954,309,508
IssuesEvent
2022-12-18 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Fri, 16 Dec 22
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Mod-Squad: Designing Mixture of Experts As Modular Multi-Task Learners - **Authors:** Zitian Chen, Yikang Shen, Mingyu Ding, Zhenfang Chen, Hengshuang Zhao, Erik Learned-Miller, Chuang Gan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2212.08066 - **Pdf link:** https://arxiv.org/pdf/2212.08066 - **Abstract** Optimization in multi-task learning (MTL) is more challenging than single-task learning (STL), as the gradient from different tasks can be contradictory. When tasks are related, it can be beneficial to share some parameters among them (cooperation). However, some tasks require additional parameters with expertise in a specific type of data or discrimination (specialization). To address the MTL challenge, we propose Mod-Squad, a new model that is Modularized into groups of experts (a 'Squad'). This structure allows us to formalize cooperation and specialization as the process of matching experts and tasks. We optimize this matching process during the training of a single model. Specifically, we incorporate mixture of experts (MoE) layers into a transformer model, with a new loss that incorporates the mutual dependence between tasks and experts. As a result, only a small set of experts are activated for each task. This prevents the sharing of the entire backbone model between all tasks, which strengthens the model, especially when the training set size and the number of tasks scale up. More interestingly, for each task, we can extract the small set of experts as a standalone model that maintains the same performance as the large model. Extensive experiments on the Taskonomy dataset with 13 vision tasks and the PASCAL-Context dataset with 5 vision tasks show the superiority of our approach. ## Keyword: event camera ### Event-based Visual Tracking in Dynamic Environments - **Authors:** Irene Perez-Salesa, Rodrigo Aldana-Lopez, Carlos Sagues - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.07754 - **Pdf link:** https://arxiv.org/pdf/2212.07754 - **Abstract** Visual object tracking under challenging conditions of motion and light can be hindered by the capabilities of conventional cameras, prone to producing images with motion blur. Event cameras are novel sensors suited to robustly perform vision tasks under these conditions. However, due to the nature of their output, applying them to object detection and tracking is non-trivial. In this work, we propose a framework to take advantage of both event cameras and off-the-shelf deep learning for object tracking. We show that reconstructing event data into intensity frames improves the tracking performance in conditions under which conventional cameras fail to provide acceptable results. ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Enhanced Training of Query-Based Object Detection via Selective Query Recollection - **Authors:** Fangyi Chen, Han Zhang, Kai Hu, Yu-kai Huang, Chenchen Zhu, Marios Savvides - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.07593 - **Pdf link:** https://arxiv.org/pdf/2212.07593 - **Abstract** This paper investigates a phenomenon where query-based object detectors mispredict at the last decoding stage while predicting correctly at an intermediate stage. We review the training process and attribute the overlooked phenomenon to two limitations: lack of training emphasis and cascading errors from decoding sequence. We design and present Selective Query Recollection (SQR), a simple and effective training strategy for query-based object detectors. It cumulatively collects intermediate queries as decoding stages go deeper and selectively forwards the queries to the downstream stages aside from the sequential structure. Such-wise, SQR places training emphasis on later stages and allows later stages to work with intermediate queries from earlier stages directly. SQR can be easily plugged into various query-based object detectors and significantly enhances their performance while leaving the inference pipeline unchanged. As a result, we apply SQR on Adamixer, DAB-DETR, and Deformable-DETR across various settings (backbone, number of queries, schedule) and consistently brings 1.4-2.8 AP improvement. ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: image signal processing There is no result ## Keyword: image signal process ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: compression ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: RAW ### Learning to Detect Semantic Boundaries with Image-level Class Labels - **Authors:** Namyup Kim, Sehyun Hwang, Suha Kwak - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2212.07579 - **Pdf link:** https://arxiv.org/pdf/2212.07579 - **Abstract** This paper presents the first attempt to learn semantic boundary detection using image-level class labels as supervision. Our method starts by estimating coarse areas of object classes through attentions drawn by an image classification network. Since boundaries will locate somewhere between such areas of different classes, our task is formulated as a multiple instance learning (MIL) problem, where pixels on a line segment connecting areas of two different classes are regarded as a bag of boundary candidates. Moreover, we design a new neural network architecture that can learn to estimate semantic boundaries reliably even with uncertain supervision given by the MIL strategy. Our network is used to generate pseudo semantic boundary labels of training images, which are in turn used to train fully supervised models. The final model trained with our pseudo labels achieves an outstanding performance on the SBD dataset, where it is as competitive as some of previous arts trained with stronger supervision. ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: raw image ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times.
2.0
New submissions for Fri, 16 Dec 22 - ## Keyword: events ### Mod-Squad: Designing Mixture of Experts As Modular Multi-Task Learners - **Authors:** Zitian Chen, Yikang Shen, Mingyu Ding, Zhenfang Chen, Hengshuang Zhao, Erik Learned-Miller, Chuang Gan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2212.08066 - **Pdf link:** https://arxiv.org/pdf/2212.08066 - **Abstract** Optimization in multi-task learning (MTL) is more challenging than single-task learning (STL), as the gradient from different tasks can be contradictory. When tasks are related, it can be beneficial to share some parameters among them (cooperation). However, some tasks require additional parameters with expertise in a specific type of data or discrimination (specialization). To address the MTL challenge, we propose Mod-Squad, a new model that is Modularized into groups of experts (a 'Squad'). This structure allows us to formalize cooperation and specialization as the process of matching experts and tasks. We optimize this matching process during the training of a single model. Specifically, we incorporate mixture of experts (MoE) layers into a transformer model, with a new loss that incorporates the mutual dependence between tasks and experts. As a result, only a small set of experts are activated for each task. This prevents the sharing of the entire backbone model between all tasks, which strengthens the model, especially when the training set size and the number of tasks scale up. More interestingly, for each task, we can extract the small set of experts as a standalone model that maintains the same performance as the large model. Extensive experiments on the Taskonomy dataset with 13 vision tasks and the PASCAL-Context dataset with 5 vision tasks show the superiority of our approach. ## Keyword: event camera ### Event-based Visual Tracking in Dynamic Environments - **Authors:** Irene Perez-Salesa, Rodrigo Aldana-Lopez, Carlos Sagues - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.07754 - **Pdf link:** https://arxiv.org/pdf/2212.07754 - **Abstract** Visual object tracking under challenging conditions of motion and light can be hindered by the capabilities of conventional cameras, prone to producing images with motion blur. Event cameras are novel sensors suited to robustly perform vision tasks under these conditions. However, due to the nature of their output, applying them to object detection and tracking is non-trivial. In this work, we propose a framework to take advantage of both event cameras and off-the-shelf deep learning for object tracking. We show that reconstructing event data into intensity frames improves the tracking performance in conditions under which conventional cameras fail to provide acceptable results. ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Enhanced Training of Query-Based Object Detection via Selective Query Recollection - **Authors:** Fangyi Chen, Han Zhang, Kai Hu, Yu-kai Huang, Chenchen Zhu, Marios Savvides - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.07593 - **Pdf link:** https://arxiv.org/pdf/2212.07593 - **Abstract** This paper investigates a phenomenon where query-based object detectors mispredict at the last decoding stage while predicting correctly at an intermediate stage. We review the training process and attribute the overlooked phenomenon to two limitations: lack of training emphasis and cascading errors from decoding sequence. We design and present Selective Query Recollection (SQR), a simple and effective training strategy for query-based object detectors. It cumulatively collects intermediate queries as decoding stages go deeper and selectively forwards the queries to the downstream stages aside from the sequential structure. Such-wise, SQR places training emphasis on later stages and allows later stages to work with intermediate queries from earlier stages directly. SQR can be easily plugged into various query-based object detectors and significantly enhances their performance while leaving the inference pipeline unchanged. As a result, we apply SQR on Adamixer, DAB-DETR, and Deformable-DETR across various settings (backbone, number of queries, schedule) and consistently brings 1.4-2.8 AP improvement. ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: image signal processing There is no result ## Keyword: image signal process ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: compression ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: RAW ### Learning to Detect Semantic Boundaries with Image-level Class Labels - **Authors:** Namyup Kim, Sehyun Hwang, Suha Kwak - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2212.07579 - **Pdf link:** https://arxiv.org/pdf/2212.07579 - **Abstract** This paper presents the first attempt to learn semantic boundary detection using image-level class labels as supervision. Our method starts by estimating coarse areas of object classes through attentions drawn by an image classification network. Since boundaries will locate somewhere between such areas of different classes, our task is formulated as a multiple instance learning (MIL) problem, where pixels on a line segment connecting areas of two different classes are regarded as a bag of boundary candidates. Moreover, we design a new neural network architecture that can learn to estimate semantic boundaries reliably even with uncertain supervision given by the MIL strategy. Our network is used to generate pseudo semantic boundary labels of training images, which are in turn used to train fully supervised models. The final model trained with our pseudo labels achieves an outstanding performance on the SBD dataset, where it is as competitive as some of previous arts trained with stronger supervision. ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times. ## Keyword: raw image ### Efficient Visual Computing with Camera RAW Snapshots - **Authors:** Zhihao Li, Ming Lu, Xu Zhang, Xin Feng, M. Salman Asif, Zhan Ma - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2212.07778 - **Pdf link:** https://arxiv.org/pdf/2212.07778 - **Abstract** Conventional cameras capture image irradiance on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel $\rho$-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained for the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on snapshots from various cameras. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression compared to RGB-domain processing. Furthermore, the proposed \r{ho}-Vision generalizes across various camera sensors and different task-specific models. Additional advantages of the proposed $\rho$-Vision that eliminates the ISP are the potential reductions in computations and processing times.
process
new submissions for fri dec keyword events mod squad designing mixture of experts as modular multi task learners authors zitian chen yikang shen mingyu ding zhenfang chen hengshuang zhao erik learned miller chuang gan subjects computer vision and pattern recognition cs cv artificial intelligence cs ai machine learning cs lg arxiv link pdf link abstract optimization in multi task learning mtl is more challenging than single task learning stl as the gradient from different tasks can be contradictory when tasks are related it can be beneficial to share some parameters among them cooperation however some tasks require additional parameters with expertise in a specific type of data or discrimination specialization to address the mtl challenge we propose mod squad a new model that is modularized into groups of experts a squad this structure allows us to formalize cooperation and specialization as the process of matching experts and tasks we optimize this matching process during the training of a single model specifically we incorporate mixture of experts moe layers into a transformer model with a new loss that incorporates the mutual dependence between tasks and experts as a result only a small set of experts are activated for each task this prevents the sharing of the entire backbone model between all tasks which strengthens the model especially when the training set size and the number of tasks scale up more interestingly for each task we can extract the small set of experts as a standalone model that maintains the same performance as the large model extensive experiments on the taskonomy dataset with vision tasks and the pascal context dataset with vision tasks show the superiority of our approach keyword event camera event based visual tracking in dynamic environments authors irene perez salesa rodrigo aldana lopez carlos sagues subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract visual object tracking under challenging conditions of motion and light can be hindered by the capabilities of conventional cameras prone to producing images with motion blur event cameras are novel sensors suited to robustly perform vision tasks under these conditions however due to the nature of their output applying them to object detection and tracking is non trivial in this work we propose a framework to take advantage of both event cameras and off the shelf deep learning for object tracking we show that reconstructing event data into intensity frames improves the tracking performance in conditions under which conventional cameras fail to provide acceptable results keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp enhanced training of query based object detection via selective query recollection authors fangyi chen han zhang kai hu yu kai huang chenchen zhu marios savvides subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract this paper investigates a phenomenon where query based object detectors mispredict at the last decoding stage while predicting correctly at an intermediate stage we review the training process and attribute the overlooked phenomenon to two limitations lack of training emphasis and cascading errors from decoding sequence we design and present selective query recollection sqr a simple and effective training strategy for query based object detectors it cumulatively collects intermediate queries as decoding stages go deeper and selectively forwards the queries to the downstream stages aside from the sequential structure such wise sqr places training emphasis on later stages and allows later stages to work with intermediate queries from earlier stages directly sqr can be easily plugged into various query based object detectors and significantly enhances their performance while leaving the inference pipeline unchanged as a result we apply sqr on adamixer dab detr and deformable detr across various settings backbone number of queries schedule and consistently brings ap improvement efficient visual computing with camera raw snapshots authors zhihao li ming lu xu zhang xin feng m salman asif zhan ma subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract conventional cameras capture image irradiance on a sensor and convert it to rgb images using an image signal processor isp the images can then be used for photography or visual computing tasks in a variety of applications such as public safety surveillance and autonomous driving one can argue that since raw images contain all the captured information the conversion of raw to rgb using an isp is not necessary for visual computing in this paper we propose a novel rho vision framework to perform high level semantic understanding and low level compression using raw images without the isp subsystem used for decades considering the scarcity of available raw image datasets we first develop an unpaired network based on unsupervised cyclegan to train modular unrolled isp and inverse isp invisp models using unpaired raw and rgb images we can then flexibly generate simulated raw images simraw using any existing rgb image dataset and finetune different models originally trained for the rgb domain to process real world camera raw images we demonstrate object detection and image compression capabilities in raw domain using raw domain and raw image compressor ric on snapshots from various cameras quantitative results reveal that raw domain task inference provides better detection accuracy and compression compared to rgb domain processing furthermore the proposed r ho vision generalizes across various camera sensors and different task specific models additional advantages of the proposed rho vision that eliminates the isp are the potential reductions in computations and processing times keyword image signal processing there is no result keyword image signal process efficient visual computing with camera raw snapshots authors zhihao li ming lu xu zhang xin feng m salman asif zhan ma subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract conventional cameras capture image irradiance on a sensor and convert it to rgb images using an image signal processor isp the images can then be used for photography or visual computing tasks in a variety of applications such as public safety surveillance and autonomous driving one can argue that since raw images contain all the captured information the conversion of raw to rgb using an isp is not necessary for visual computing in this paper we propose a novel rho vision framework to perform high level semantic understanding and low level compression using raw images without the isp subsystem used for decades considering the scarcity of available raw image datasets we first develop an unpaired network based on unsupervised cyclegan to train modular unrolled isp and inverse isp invisp models using unpaired raw and rgb images we can then flexibly generate simulated raw images simraw using any existing rgb image dataset and finetune different models originally trained for the rgb domain to process real world camera raw images we demonstrate object detection and image compression capabilities in raw domain using raw domain and raw image compressor ric on snapshots from various cameras quantitative results reveal that raw domain task inference provides better detection accuracy and compression compared to rgb domain processing furthermore the proposed r ho vision generalizes across various camera sensors and different task specific models additional advantages of the proposed rho vision that eliminates the isp are the potential reductions in computations and processing times keyword compression efficient visual computing with camera raw snapshots authors zhihao li ming lu xu zhang xin feng m salman asif zhan ma subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract conventional cameras capture image irradiance on a sensor and convert it to rgb images using an image signal processor isp the images can then be used for photography or visual computing tasks in a variety of applications such as public safety surveillance and autonomous driving one can argue that since raw images contain all the captured information the conversion of raw to rgb using an isp is not necessary for visual computing in this paper we propose a novel rho vision framework to perform high level semantic understanding and low level compression using raw images without the isp subsystem used for decades considering the scarcity of available raw image datasets we first develop an unpaired network based on unsupervised cyclegan to train modular unrolled isp and inverse isp invisp models using unpaired raw and rgb images we can then flexibly generate simulated raw images simraw using any existing rgb image dataset and finetune different models originally trained for the rgb domain to process real world camera raw images we demonstrate object detection and image compression capabilities in raw domain using raw domain and raw image compressor ric on snapshots from various cameras quantitative results reveal that raw domain task inference provides better detection accuracy and compression compared to rgb domain processing furthermore the proposed r ho vision generalizes across various camera sensors and different task specific models additional advantages of the proposed rho vision that eliminates the isp are the potential reductions in computations and processing times keyword raw learning to detect semantic boundaries with image level class labels authors namyup kim sehyun hwang suha kwak subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract this paper presents the first attempt to learn semantic boundary detection using image level class labels as supervision our method starts by estimating coarse areas of object classes through attentions drawn by an image classification network since boundaries will locate somewhere between such areas of different classes our task is formulated as a multiple instance learning mil problem where pixels on a line segment connecting areas of two different classes are regarded as a bag of boundary candidates moreover we design a new neural network architecture that can learn to estimate semantic boundaries reliably even with uncertain supervision given by the mil strategy our network is used to generate pseudo semantic boundary labels of training images which are in turn used to train fully supervised models the final model trained with our pseudo labels achieves an outstanding performance on the sbd dataset where it is as competitive as some of previous arts trained with stronger supervision efficient visual computing with camera raw snapshots authors zhihao li ming lu xu zhang xin feng m salman asif zhan ma subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract conventional cameras capture image irradiance on a sensor and convert it to rgb images using an image signal processor isp the images can then be used for photography or visual computing tasks in a variety of applications such as public safety surveillance and autonomous driving one can argue that since raw images contain all the captured information the conversion of raw to rgb using an isp is not necessary for visual computing in this paper we propose a novel rho vision framework to perform high level semantic understanding and low level compression using raw images without the isp subsystem used for decades considering the scarcity of available raw image datasets we first develop an unpaired network based on unsupervised cyclegan to train modular unrolled isp and inverse isp invisp models using unpaired raw and rgb images we can then flexibly generate simulated raw images simraw using any existing rgb image dataset and finetune different models originally trained for the rgb domain to process real world camera raw images we demonstrate object detection and image compression capabilities in raw domain using raw domain and raw image compressor ric on snapshots from various cameras quantitative results reveal that raw domain task inference provides better detection accuracy and compression compared to rgb domain processing furthermore the proposed r ho vision generalizes across various camera sensors and different task specific models additional advantages of the proposed rho vision that eliminates the isp are the potential reductions in computations and processing times keyword raw image efficient visual computing with camera raw snapshots authors zhihao li ming lu xu zhang xin feng m salman asif zhan ma subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract conventional cameras capture image irradiance on a sensor and convert it to rgb images using an image signal processor isp the images can then be used for photography or visual computing tasks in a variety of applications such as public safety surveillance and autonomous driving one can argue that since raw images contain all the captured information the conversion of raw to rgb using an isp is not necessary for visual computing in this paper we propose a novel rho vision framework to perform high level semantic understanding and low level compression using raw images without the isp subsystem used for decades considering the scarcity of available raw image datasets we first develop an unpaired network based on unsupervised cyclegan to train modular unrolled isp and inverse isp invisp models using unpaired raw and rgb images we can then flexibly generate simulated raw images simraw using any existing rgb image dataset and finetune different models originally trained for the rgb domain to process real world camera raw images we demonstrate object detection and image compression capabilities in raw domain using raw domain and raw image compressor ric on snapshots from various cameras quantitative results reveal that raw domain task inference provides better detection accuracy and compression compared to rgb domain processing furthermore the proposed r ho vision generalizes across various camera sensors and different task specific models additional advantages of the proposed rho vision that eliminates the isp are the potential reductions in computations and processing times
1
21,463
29,499,897,840
IssuesEvent
2023-06-02 20:31:54
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Need functions like "expressionableColumns" (name?)
.Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
Not sure about the name. We need functions that can tell you what columns are available to be used in expressions (the custom expression editor), similar to how we have `orderableColumns` and the like. I'm not sure what other functions we might need to power the UI, let's go thru it and try to work out what we'll need to be able to power it. Note that implementing the custom expression parser is not on our plate for the time being
1.0
[MLv2] Need functions like "expressionableColumns" (name?) - Not sure about the name. We need functions that can tell you what columns are available to be used in expressions (the custom expression editor), similar to how we have `orderableColumns` and the like. I'm not sure what other functions we might need to power the UI, let's go thru it and try to work out what we'll need to be able to power it. Note that implementing the custom expression parser is not on our plate for the time being
process
need functions like expressionablecolumns name not sure about the name we need functions that can tell you what columns are available to be used in expressions the custom expression editor similar to how we have orderablecolumns and the like i m not sure what other functions we might need to power the ui let s go thru it and try to work out what we ll need to be able to power it note that implementing the custom expression parser is not on our plate for the time being
1