Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
8,705
11,847,053,507
IssuesEvent
2020-03-24 11:18:04
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
68000/big/32/MC68020 decompile fail
Feature: Processor/68000 Type: Bug
**Describe the bug** There is a really simple function and the tool just translates it to "return" **To Reproduce** Given the following 68K ``` ************************************************************** * FUNCTION * ************************************************************** undefined getDistaceD1_2toD5_6inD7() undefined D0b:1 <RETURN> undefined2 Stack[-0x8]:2 local_8 XREF[1]: 0002098e(R) getDistaceD1_2toD5_6inD7 XREF[1]: FUN_00020994:000209a2(c) 0002097a 48 a7 60 00 movem.w { D2w D1w},-(SP) 0002097e 94 46 sub.w D6w,D2w 00020980 6c 02 bge.b LAB_00020984 00020982 44 42 neg.w D2w LAB_00020984 XREF[1]: 00020980(j) 00020984 92 45 sub.w D5w,D1w 00020986 6c 02 bge.b LAB_0002098a 00020988 44 41 neg.w D1w LAB_0002098a XREF[1]: 00020986(j) 0002098a d2 42 add.w D2w,D1w 0002098c 3e 01 move.w D1w,D7w 0002098e 4c 9f 00 06 movem.w (SP=>local_8)+,{ D1 D2} 00020992 4e 75 rts ``` and it gives you ``` void getDistaceD1_2toD5_6inD7(void) { return; } ``` **Expected behavior** this function is clearly should be something along the lines of ``` D2 = abs((short)D6-(short)D2); D1 = abs((short)D5-(short)D1); D7 = D1+D2; return D7; ``` **Screenshots** If applicable, add screenshots to help explain your problem. **Attachments** I've attached the xml log from the "Debug Function Compilation" button in the Decompile Window. [68K0002097a.zip](https://github.com/NationalSecurityAgency/ghidra/files/4317864/68K0002097a.zip) **Environment (please complete the following information):** - OS: Windows 10.0.18363 - Java Version: 11.0.6.10.1-1 corretto - Ghidra Version: 9.1.2_PUBLIC_20200212 **Additional context** I have this plug in installed which I patched to work on 9.1.2 by changing the version number in the properties file. https://github.com/lab313ru/ghidra_sega_ldr
1.0
68000/big/32/MC68020 decompile fail - **Describe the bug** There is a really simple function and the tool just translates it to "return" **To Reproduce** Given the following 68K ``` ************************************************************** * FUNCTION * ************************************************************** undefined getDistaceD1_2toD5_6inD7() undefined D0b:1 <RETURN> undefined2 Stack[-0x8]:2 local_8 XREF[1]: 0002098e(R) getDistaceD1_2toD5_6inD7 XREF[1]: FUN_00020994:000209a2(c) 0002097a 48 a7 60 00 movem.w { D2w D1w},-(SP) 0002097e 94 46 sub.w D6w,D2w 00020980 6c 02 bge.b LAB_00020984 00020982 44 42 neg.w D2w LAB_00020984 XREF[1]: 00020980(j) 00020984 92 45 sub.w D5w,D1w 00020986 6c 02 bge.b LAB_0002098a 00020988 44 41 neg.w D1w LAB_0002098a XREF[1]: 00020986(j) 0002098a d2 42 add.w D2w,D1w 0002098c 3e 01 move.w D1w,D7w 0002098e 4c 9f 00 06 movem.w (SP=>local_8)+,{ D1 D2} 00020992 4e 75 rts ``` and it gives you ``` void getDistaceD1_2toD5_6inD7(void) { return; } ``` **Expected behavior** this function is clearly should be something along the lines of ``` D2 = abs((short)D6-(short)D2); D1 = abs((short)D5-(short)D1); D7 = D1+D2; return D7; ``` **Screenshots** If applicable, add screenshots to help explain your problem. **Attachments** I've attached the xml log from the "Debug Function Compilation" button in the Decompile Window. [68K0002097a.zip](https://github.com/NationalSecurityAgency/ghidra/files/4317864/68K0002097a.zip) **Environment (please complete the following information):** - OS: Windows 10.0.18363 - Java Version: 11.0.6.10.1-1 corretto - Ghidra Version: 9.1.2_PUBLIC_20200212 **Additional context** I have this plug in installed which I patched to work on 9.1.2 by changing the version number in the properties file. https://github.com/lab313ru/ghidra_sega_ldr
process
big decompile fail describe the bug there is a really simple function and the tool just translates it to return to reproduce given the following function undefined undefined stack local xref r xref fun c movem w sp sub w bge b lab neg w lab xref j sub w bge b lab neg w lab xref j add w move w movem w sp local rts and it gives you void void return expected behavior this function is clearly should be something along the lines of abs short short abs short short return screenshots if applicable add screenshots to help explain your problem attachments i ve attached the xml log from the debug function compilation button in the decompile window environment please complete the following information os windows java version corretto ghidra version public additional context i have this plug in installed which i patched to work on by changing the version number in the properties file
1
722,362
24,860,532,831
IssuesEvent
2022-10-27 07:54:42
AdamCvijanovic/Herding-Ghosts-Unity
https://api.github.com/repos/AdamCvijanovic/Herding-Ghosts-Unity
closed
Introduction to the Scene
Enhance/Important for Gameplay High Priority
Introduction by Graddaughter/narrator to give the player clarification on the story/settings/objectives etc.
1.0
Introduction to the Scene - Introduction by Graddaughter/narrator to give the player clarification on the story/settings/objectives etc.
non_process
introduction to the scene introduction by graddaughter narrator to give the player clarification on the story settings objectives etc
0
65,851
3,244,494,082
IssuesEvent
2015-10-16 02:45:21
cs2103aug2015-f10-3j/main
https://api.github.com/repos/cs2103aug2015-f10-3j/main
opened
Edit Support for reminder
priority.high type.task
edit of deadline will modify the reminder respectively 1. calculate the difference between the old deadline and reminder 2. apply the difference on the new deadline and reminder edit of reminder directly - edit 1 remind dd/mm/yyyy hh:mm - edit 1 -r dd/mm/yyyy hh:mm
1.0
Edit Support for reminder - edit of deadline will modify the reminder respectively 1. calculate the difference between the old deadline and reminder 2. apply the difference on the new deadline and reminder edit of reminder directly - edit 1 remind dd/mm/yyyy hh:mm - edit 1 -r dd/mm/yyyy hh:mm
non_process
edit support for reminder edit of deadline will modify the reminder respectively calculate the difference between the old deadline and reminder apply the difference on the new deadline and reminder edit of reminder directly edit remind dd mm yyyy hh mm edit r dd mm yyyy hh mm
0
725,561
24,966,312,940
IssuesEvent
2022-11-01 19:42:52
dtcenter/METviewer
https://api.github.com/repos/dtcenter/METviewer
opened
Plot the new ECNT statistics: MAE, MAE_OERR, BIAS_RATIO, ME_LT_OBS, and ME_GT_OBS
type: enhancement priority: high requestor: NOAA/EMC alert: NEED ACCOUNT KEY requestor: UK Met Office METviewer: Plotting METviewer: Statistics METviewer: Scorecard
## Describe the Enhancement ## MET development version 11.0.0-beta4 made several changes to the "stat" output. This issue is to enhance METviewer to plot the newly added statistics and enable them to be included in scorecards. I don't see a need to plot the N_GE_OBS and N_LT_OBS columns that were also added, but they are needed in the aggregation of the other columns. - Issue dtcenter/MET#2325 adds 2 new columns (MAE and MAE_OERR) to the ECNT line type generated by Ensemble-Stat and the HiRA method in Point-Stat. See [Table 13.2](https://met.readthedocs.io/en/develop/Users_Guide/ensemble-stat.html?highlight=ECNT#id2). The MAE is the mean absolute error. These columns should be aggregated together over multiple cases using a weighted average approach, where the weight is defined by the number of matched pairs in the TOTAL column. These MAE's should be added to the list of statistics where 0 is the perfect score. - Issue dtcenter/MET#2058 adds 5 new columns (BIAS_RATIO, N_GE_OBS, ME_GE_OBS, N_LT_OBS, ME_LT_OBS) to the ECNT line type generated by Ensemble-Stat and the HiRA method in Point-Stat. See [Table 13.2](https://met.readthedocs.io/en/develop/Users_Guide/ensemble-stat.html?highlight=ECNT#id2). All 5 columns are added so that the BIAS_RATIO can be aggregated over multiple runs, as described below. - N_GE_OBS is an integer and is aggregated over multiple runs by summing the values. - ME_GE_OBS is a float and is aggregated as a weighted average where the weights are defined by the N_GE_OBS column. - N_LT_OBS is an integer and is aggregated over multiple runs by summing the values. - ME_LT_OBS is a float and is aggregated as a weighted average where the weights are defined by the N_LT_OBS column. - BIAS_RATIO is a float >0. It aggregated by first aggregating ME_GE_OBS and ME_LT_OBS and then computing BIAS_RATIO = ME_GE_OBS / absolute value of ME_LT_OBS. The ME_GE_OBS and ME_LT_OBS columns are mean errors and should added to list of statistics where 0 is the perfect score. The BIAS_RATIO is the ratio of those 2 ME values, and I suppose a perfect score would be 0/0 which is undefined. This *probably* should go in the list of statistics where 1 is the perfect score... where the amount of high bias equal the amount of low bias. In this way, its similar to the frequency bias (FBIAS) statistic. Sample data for this issue can be found in the nightly build output in seneca:/d1/projects/MET/MET_regression/develop/NB20221101/MET-develop/test_output (or a more recent date). And here's a tarfile of that output: ftp://ftp.rap.ucar.edu/incoming/irap/met_help/NB20221101_develop_test_output.tar.gz ### Time Estimate ### *Estimate the amount of work required here.* *Issues should represent approximately 1 to 3 days of work.* ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Enhancement Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
1.0
Plot the new ECNT statistics: MAE, MAE_OERR, BIAS_RATIO, ME_LT_OBS, and ME_GT_OBS - ## Describe the Enhancement ## MET development version 11.0.0-beta4 made several changes to the "stat" output. This issue is to enhance METviewer to plot the newly added statistics and enable them to be included in scorecards. I don't see a need to plot the N_GE_OBS and N_LT_OBS columns that were also added, but they are needed in the aggregation of the other columns. - Issue dtcenter/MET#2325 adds 2 new columns (MAE and MAE_OERR) to the ECNT line type generated by Ensemble-Stat and the HiRA method in Point-Stat. See [Table 13.2](https://met.readthedocs.io/en/develop/Users_Guide/ensemble-stat.html?highlight=ECNT#id2). The MAE is the mean absolute error. These columns should be aggregated together over multiple cases using a weighted average approach, where the weight is defined by the number of matched pairs in the TOTAL column. These MAE's should be added to the list of statistics where 0 is the perfect score. - Issue dtcenter/MET#2058 adds 5 new columns (BIAS_RATIO, N_GE_OBS, ME_GE_OBS, N_LT_OBS, ME_LT_OBS) to the ECNT line type generated by Ensemble-Stat and the HiRA method in Point-Stat. See [Table 13.2](https://met.readthedocs.io/en/develop/Users_Guide/ensemble-stat.html?highlight=ECNT#id2). All 5 columns are added so that the BIAS_RATIO can be aggregated over multiple runs, as described below. - N_GE_OBS is an integer and is aggregated over multiple runs by summing the values. - ME_GE_OBS is a float and is aggregated as a weighted average where the weights are defined by the N_GE_OBS column. - N_LT_OBS is an integer and is aggregated over multiple runs by summing the values. - ME_LT_OBS is a float and is aggregated as a weighted average where the weights are defined by the N_LT_OBS column. - BIAS_RATIO is a float >0. It aggregated by first aggregating ME_GE_OBS and ME_LT_OBS and then computing BIAS_RATIO = ME_GE_OBS / absolute value of ME_LT_OBS. The ME_GE_OBS and ME_LT_OBS columns are mean errors and should added to list of statistics where 0 is the perfect score. The BIAS_RATIO is the ratio of those 2 ME values, and I suppose a perfect score would be 0/0 which is undefined. This *probably* should go in the list of statistics where 1 is the perfect score... where the amount of high bias equal the amount of low bias. In this way, its similar to the frequency bias (FBIAS) statistic. Sample data for this issue can be found in the nightly build output in seneca:/d1/projects/MET/MET_regression/develop/NB20221101/MET-develop/test_output (or a more recent date). And here's a tarfile of that output: ftp://ftp.rap.ucar.edu/incoming/irap/met_help/NB20221101_develop_test_output.tar.gz ### Time Estimate ### *Estimate the amount of work required here.* *Issues should represent approximately 1 to 3 days of work.* ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Enhancement Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
non_process
plot the new ecnt statistics mae mae oerr bias ratio me lt obs and me gt obs describe the enhancement met development version made several changes to the stat output this issue is to enhance metviewer to plot the newly added statistics and enable them to be included in scorecards i don t see a need to plot the n ge obs and n lt obs columns that were also added but they are needed in the aggregation of the other columns issue dtcenter met adds new columns mae and mae oerr to the ecnt line type generated by ensemble stat and the hira method in point stat see the mae is the mean absolute error these columns should be aggregated together over multiple cases using a weighted average approach where the weight is defined by the number of matched pairs in the total column these mae s should be added to the list of statistics where is the perfect score issue dtcenter met adds new columns bias ratio n ge obs me ge obs n lt obs me lt obs to the ecnt line type generated by ensemble stat and the hira method in point stat see all columns are added so that the bias ratio can be aggregated over multiple runs as described below n ge obs is an integer and is aggregated over multiple runs by summing the values me ge obs is a float and is aggregated as a weighted average where the weights are defined by the n ge obs column n lt obs is an integer and is aggregated over multiple runs by summing the values me lt obs is a float and is aggregated as a weighted average where the weights are defined by the n lt obs column bias ratio is a float it aggregated by first aggregating me ge obs and me lt obs and then computing bias ratio me ge obs absolute value of me lt obs the me ge obs and me lt obs columns are mean errors and should added to list of statistics where is the perfect score the bias ratio is the ratio of those me values and i suppose a perfect score would be which is undefined this probably should go in the list of statistics where is the perfect score where the amount of high bias equal the amount of low bias in this way its similar to the frequency bias fbias statistic sample data for this issue can be found in the nightly build output in seneca projects met met regression develop met develop test output or a more recent date and here s a tarfile of that output ftp ftp rap ucar edu incoming irap met help develop test output tar gz time estimate estimate the amount of work required here issues should represent approximately to days of work sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
0
108,405
9,305,768,904
IssuesEvent
2019-03-25 07:46:37
oscoin/oscoin
https://api.github.com/repos/oscoin/oscoin
closed
Investigate fork selection equivalence mismatch
testing
From this [build](https://console.cloud.google.com/cloud-build/builds/c671143e-2055-4d53-b662-d392c22dc782?project=opensourcecoin): ``` Storage.Block SQLite Storage blackbox Use --quickcheck-replay=917375 to reproduce. Block {blockHeader = BlockHeader {blockPrevHash = "R1ELJ1v", blockDataHash = "Pji2uFs", blockStateHash = "4J2q5RT", blockTimestamp = Timestamp 42390614099162, blockTargetDifficulty = Difficulty {fromDifficulty = 0}, blockSeal = SealedWith "</"}, blockHash = "YkoZBET", blockData = fromList [Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-1852296587708) % 3229532965117))), sigSignature = MockSignature [94,94,219,71,241,245,106,213]}, txPubKey = MockPK (PK (MockKey [94,94,219,71,241,245,106,213]) (Hashed "hFqXpv6")), txChainId = 1, txNonce = 2, txContext = "8mni7zR"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (5222099247581 % 3418035084665))), sigSignature = MockSignature [29,14,55,93,211,139,73,7]}, txPubKey = MockPK (PK (MockKey [29,14,55,93,211,139,73,7]) (Hashed "JQ1cYrs")), txChainId = 4, txNonce = 2, txContext = "PeFdHk1"}]} /= Block {blockHeader = BlockHeader {blockPrevHash = "fiu6biF", blockDataHash = "CQZouvr", blockStateHash = "W3sdJFe", blockTimestamp = Timestamp 42831309040803, blockTargetDifficulty = Difficulty {fromDifficulty = 0}, blockSeal = SealedWith "\FS\727784;\ESCk\697266H\FSJ>a9"}, blockHash = "ZnqcE1e", blockData = fromList [Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (1977492603782 % 288807972433))), sigSignature = MockSignature [73,38,184,8,190,126,136,60]}, txPubKey = MockPK (PK (MockKey [73,38,184,8,190,126,136,60]) (Hashed "iGEWgEr")), txChainId = 98, txNonce = 109, txContext = "Yk5tGf7"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-28201519195159) % 3736384581802))), sigSignature = MockSignature [6,125,56,159,174,249,10,139]}, txPubKey = MockPK (PK (MockKey [6,125,56,159,174,249,10,139]) (Hashed "cLA7EJa")), txChainId = 111, txNonce = 6, txContext = "E1jHfyJ"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-355532849589) % 42963497534))), sigSignature = MockSignature [128,212,135,150,69,28,113,252]}, txPubKey = MockPK (PK (MockKey [128,212,135,150,69,28,113,252]) (Hashed "17b3Ar4")), txChainId = 96, txNonce = 12, txContext = "KGXECx3"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-25767441461675) % 4030487530446))), sigSignature = MockSignature [70,125,36,162,72,82,30,88]}, txPubKey = MockPK (PK (MockKey [70,125,36,162,72,82,30,88]) (Hashed "23k7oFi")), txChainId = 45, txNonce = 32, txContext = "GACiqgJ"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (60276902944458 % 6914840929649))), sigSignature = MockSignature [102,170,127,84,101,236,157,183]}, txPubKey = MockPK (PK (MockKey [102,170,127,84,101,236,157,183]) (Hashed "bnGW4UW")), txChainId = 23, txNonce = 53, txContext = "5nEfX4c"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (137286829419554 % 9981493660417))), sigSignature = MockSignature [87,249,150,172,214,33,123,39]}, txPubKey = MockPK (PK (MockKey [87,249,150,172,214,33,123,39]) (Hashed "18Uj25r")), txChainId = 2, txNonce = 56, txContext = "LatecBL"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (BooleanF True)), sigSignature = MockSignature [239,85,164,55,35,110,253,211]}, txPubKey = MockPK (PK (MockKey [239,85,164,55,35,110,253,211]) (Hashed "MFP3qjd")), txChainId = 39, txNonce = 37, txContext = "W7Q1Kbi"}]} Counterexample is: pureStore = [ZnqcE1e (p:fiu6biF) @11h] sqlStore = [YkoZBET (p:R1ELJ1v) @11h] api call at test/Oscoin/Test/Storage/Block/Equivalence.hs:90:23 yielded a result mismatch! - [YhzpbCv (p:18auNLF) @1h] link: [18auNLF (p:aCrGXLe) @52m] - [CFAoYmu (p:fCLrkMk) @5h] ← [7JSLdiD (p:CFAoYmu) @6h] ← [c3doitG (p:7JSLdiD) @7h] ← [M7A3wDz (p:c3doitG) @8h] ← [TQu7CWu (p:M7A3wDz) @9h] ← [hwm9btQ (p:TQu7CWu) @10h] ← [cuq2wJM (p:hwm9btQ) @11h] ← [9HuczZR (p:cuq2wJM) @11h] link: [fCLrkMk (p:9baqMPC) @5h] - [R1ELJ1v (p:Ek5hLc3) @10h] ← [YkoZBET (p:R1ELJ1v) @11h] link: [Ek5hLc3 (p:CRdNfUZ) @10h] orphans: chain: [aCrGXLe (p:1111111) @0ns] ← [fXrENhp (p:aCrGXLe) @52m] ← [iGdkJhJ (p:fXrENhp) @1h] ← [8BnLM8q (p:iGdkJhJ) @2h] ← [Gev6P2z (p:8BnLM8q) @3h] ← [9baqMPC (p:Gev6P2z) @4h] ← [dbzPAag (p:9baqMPC) @5h] ← [cMEU73D (p:dbzPAag) @5h] ← [5o4ZcM9 (p:cMEU73D) @6h] ← [V3NTieW (p:5o4ZcM9) @7h] ← [5bfvLRu (p:V3NTieW) @8h] ← [CRdNfUZ (p:5bfvLRu) @9h] ← [NNAKYgj (p:CRdNfUZ) @10h] ← [fiu6biF (p:NNAKYgj) @11h] ← [ZnqcE1e (p:fiu6biF) @11h] *** Failed! Falsifiable (after 12 tests): (forks) getTip . insertBlock equivalence: FAIL (3.54s) ```
1.0
Investigate fork selection equivalence mismatch - From this [build](https://console.cloud.google.com/cloud-build/builds/c671143e-2055-4d53-b662-d392c22dc782?project=opensourcecoin): ``` Storage.Block SQLite Storage blackbox Use --quickcheck-replay=917375 to reproduce. Block {blockHeader = BlockHeader {blockPrevHash = "R1ELJ1v", blockDataHash = "Pji2uFs", blockStateHash = "4J2q5RT", blockTimestamp = Timestamp 42390614099162, blockTargetDifficulty = Difficulty {fromDifficulty = 0}, blockSeal = SealedWith "</"}, blockHash = "YkoZBET", blockData = fromList [Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-1852296587708) % 3229532965117))), sigSignature = MockSignature [94,94,219,71,241,245,106,213]}, txPubKey = MockPK (PK (MockKey [94,94,219,71,241,245,106,213]) (Hashed "hFqXpv6")), txChainId = 1, txNonce = 2, txContext = "8mni7zR"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (5222099247581 % 3418035084665))), sigSignature = MockSignature [29,14,55,93,211,139,73,7]}, txPubKey = MockPK (PK (MockKey [29,14,55,93,211,139,73,7]) (Hashed "JQ1cYrs")), txChainId = 4, txNonce = 2, txContext = "PeFdHk1"}]} /= Block {blockHeader = BlockHeader {blockPrevHash = "fiu6biF", blockDataHash = "CQZouvr", blockStateHash = "W3sdJFe", blockTimestamp = Timestamp 42831309040803, blockTargetDifficulty = Difficulty {fromDifficulty = 0}, blockSeal = SealedWith "\FS\727784;\ESCk\697266H\FSJ>a9"}, blockHash = "ZnqcE1e", blockData = fromList [Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (1977492603782 % 288807972433))), sigSignature = MockSignature [73,38,184,8,190,126,136,60]}, txPubKey = MockPK (PK (MockKey [73,38,184,8,190,126,136,60]) (Hashed "iGEWgEr")), txChainId = 98, txNonce = 109, txContext = "Yk5tGf7"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-28201519195159) % 3736384581802))), sigSignature = MockSignature [6,125,56,159,174,249,10,139]}, txPubKey = MockPK (PK (MockKey [6,125,56,159,174,249,10,139]) (Hashed "cLA7EJa")), txChainId = 111, txNonce = 6, txContext = "E1jHfyJ"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-355532849589) % 42963497534))), sigSignature = MockSignature [128,212,135,150,69,28,113,252]}, txPubKey = MockPK (PK (MockKey [128,212,135,150,69,28,113,252]) (Hashed "17b3Ar4")), txChainId = 96, txNonce = 12, txContext = "KGXECx3"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF ((-25767441461675) % 4030487530446))), sigSignature = MockSignature [70,125,36,162,72,82,30,88]}, txPubKey = MockPK (PK (MockKey [70,125,36,162,72,82,30,88]) (Hashed "23k7oFi")), txChainId = 45, txNonce = 32, txContext = "GACiqgJ"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (60276902944458 % 6914840929649))), sigSignature = MockSignature [102,170,127,84,101,236,157,183]}, txPubKey = MockPK (PK (MockKey [102,170,127,84,101,236,157,183]) (Hashed "bnGW4UW")), txChainId = 23, txNonce = 53, txContext = "5nEfX4c"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (NumberF (137286829419554 % 9981493660417))), sigSignature = MockSignature [87,249,150,172,214,33,123,39]}, txPubKey = MockPK (PK (MockKey [87,249,150,172,214,33,123,39]) (Hashed "18Uj25r")), txChainId = 2, txNonce = 56, txContext = "LatecBL"},Tx {txMessage = Signed {sigMessage = Annotated (WithPos (InternalPos "src/Radicle/Internal/Annotation.hs:79:28 thisPos\nsrc/Radicle/Internal/Annotation.hs:36:50 toAnnotation") (BooleanF True)), sigSignature = MockSignature [239,85,164,55,35,110,253,211]}, txPubKey = MockPK (PK (MockKey [239,85,164,55,35,110,253,211]) (Hashed "MFP3qjd")), txChainId = 39, txNonce = 37, txContext = "W7Q1Kbi"}]} Counterexample is: pureStore = [ZnqcE1e (p:fiu6biF) @11h] sqlStore = [YkoZBET (p:R1ELJ1v) @11h] api call at test/Oscoin/Test/Storage/Block/Equivalence.hs:90:23 yielded a result mismatch! - [YhzpbCv (p:18auNLF) @1h] link: [18auNLF (p:aCrGXLe) @52m] - [CFAoYmu (p:fCLrkMk) @5h] ← [7JSLdiD (p:CFAoYmu) @6h] ← [c3doitG (p:7JSLdiD) @7h] ← [M7A3wDz (p:c3doitG) @8h] ← [TQu7CWu (p:M7A3wDz) @9h] ← [hwm9btQ (p:TQu7CWu) @10h] ← [cuq2wJM (p:hwm9btQ) @11h] ← [9HuczZR (p:cuq2wJM) @11h] link: [fCLrkMk (p:9baqMPC) @5h] - [R1ELJ1v (p:Ek5hLc3) @10h] ← [YkoZBET (p:R1ELJ1v) @11h] link: [Ek5hLc3 (p:CRdNfUZ) @10h] orphans: chain: [aCrGXLe (p:1111111) @0ns] ← [fXrENhp (p:aCrGXLe) @52m] ← [iGdkJhJ (p:fXrENhp) @1h] ← [8BnLM8q (p:iGdkJhJ) @2h] ← [Gev6P2z (p:8BnLM8q) @3h] ← [9baqMPC (p:Gev6P2z) @4h] ← [dbzPAag (p:9baqMPC) @5h] ← [cMEU73D (p:dbzPAag) @5h] ← [5o4ZcM9 (p:cMEU73D) @6h] ← [V3NTieW (p:5o4ZcM9) @7h] ← [5bfvLRu (p:V3NTieW) @8h] ← [CRdNfUZ (p:5bfvLRu) @9h] ← [NNAKYgj (p:CRdNfUZ) @10h] ← [fiu6biF (p:NNAKYgj) @11h] ← [ZnqcE1e (p:fiu6biF) @11h] *** Failed! Falsifiable (after 12 tests): (forks) getTip . insertBlock equivalence: FAIL (3.54s) ```
non_process
investigate fork selection equivalence mismatch from this storage block sqlite storage blackbox use quickcheck replay to reproduce block blockheader blockheader blockprevhash blockdatahash blockstatehash blocktimestamp timestamp blocktargetdifficulty difficulty fromdifficulty blockseal sealedwith blockhash blockdata fromlist txpubkey mockpk pk mockkey hashed igewger txchainid txnonce txcontext tx txmessage signed sigmessage annotated withpos internalpos src radicle internal annotation hs thispos nsrc radicle internal annotation hs toannotation numberf sigsignature mocksignature txpubkey mockpk pk mockkey hashed txchainid txnonce txcontext tx txmessage signed sigmessage annotated withpos internalpos src radicle internal annotation hs thispos nsrc radicle internal annotation hs toannotation numberf sigsignature mocksignature txpubkey mockpk pk mockkey hashed txchainid txnonce txcontext tx txmessage signed sigmessage annotated withpos internalpos src radicle internal annotation hs thispos nsrc radicle internal annotation hs toannotation numberf sigsignature mocksignature txpubkey mockpk pk mockkey hashed txchainid txnonce txcontext gaciqgj tx txmessage signed sigmessage annotated withpos internalpos src radicle internal annotation hs thispos nsrc radicle internal annotation hs toannotation numberf sigsignature mocksignature txpubkey mockpk pk mockkey hashed txchainid txnonce txcontext tx txmessage signed sigmessage annotated withpos internalpos src radicle internal annotation hs thispos nsrc radicle internal annotation hs toannotation numberf sigsignature mocksignature txpubkey mockpk pk mockkey hashed txchainid txnonce txcontext latecbl tx txmessage signed sigmessage annotated withpos internalpos src radicle internal annotation hs thispos nsrc radicle internal annotation hs toannotation booleanf true sigsignature mocksignature txpubkey mockpk pk mockkey hashed txchainid txnonce txcontext counterexample is purestore sqlstore api call at test oscoin test storage block equivalence hs yielded a result mismatch link ← ← ← ← ← ← ← link ← link orphans chain ← ← ← ← ← ← ← ← ← ← ← ← ← ← failed falsifiable after tests forks gettip insertblock equivalence fail
0
20,972
27,819,526,712
IssuesEvent
2023-03-19 03:27:23
cse442-at-ub/project_s23-team-infinity
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
closed
Create the login backend functionality of the login page
Processing Task Sprint 2
**Task Test** *Test1* 1) Add an email/username and password pair in our SQL database 2) Go to the website that is provided (https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/) 3) Verify the page is in the login page 4) Verify there are two fields in the left for user to type their email and password 5) Type the email/username and password we created in step 1 in the fields 6) Click login button 7) Verify the page redirects to the home page *Test 2* 1) Add an email/username and password pair in our SQL database 2) Go to the website that is provided (https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/) 3) Verify the page is in the login page 4) Verify there are two fields in the left for user to type their email and password 5) Type an email/username and password that is different from what we created in step 1 in the fields 6) Click login button 7) Verify the page stay in the login page 8) Verify that a error message pops out in the login page
1.0
Create the login backend functionality of the login page - **Task Test** *Test1* 1) Add an email/username and password pair in our SQL database 2) Go to the website that is provided (https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/) 3) Verify the page is in the login page 4) Verify there are two fields in the left for user to type their email and password 5) Type the email/username and password we created in step 1 in the fields 6) Click login button 7) Verify the page redirects to the home page *Test 2* 1) Add an email/username and password pair in our SQL database 2) Go to the website that is provided (https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/) 3) Verify the page is in the login page 4) Verify there are two fields in the left for user to type their email and password 5) Type an email/username and password that is different from what we created in step 1 in the fields 6) Click login button 7) Verify the page stay in the login page 8) Verify that a error message pops out in the login page
process
create the login backend functionality of the login page task test add an email username and password pair in our sql database go to the website that is provided verify the page is in the login page verify there are two fields in the left for user to type their email and password type the email username and password we created in step in the fields click login button verify the page redirects to the home page test add an email username and password pair in our sql database go to the website that is provided verify the page is in the login page verify there are two fields in the left for user to type their email and password type an email username and password that is different from what we created in step in the fields click login button verify the page stay in the login page verify that a error message pops out in the login page
1
7,188
10,328,630,437
IssuesEvent
2019-09-02 09:59:43
lyc10031/gitalk
https://api.github.com/repos/lyc10031/gitalk
opened
gin框架的调用流程 - Ray_YC's Blog
Gitalk gin-processes
https://lyc10031.github.io/2019/09/02/gin%E8%B0%83%E7%94%A8%E6%B5%81%E7%A8%8B.html 文件树:my-gin1/├── conf│   └── app.ini├── main.go├── middleware├── models│   ├── models.go│   └── tag.go├── pkg│   ├── e│   │   ├── code.go│   │   └── msg.go│  ...
1.0
gin框架的调用流程 - Ray_YC's Blog - https://lyc10031.github.io/2019/09/02/gin%E8%B0%83%E7%94%A8%E6%B5%81%E7%A8%8B.html 文件树:my-gin1/├── conf│   └── app.ini├── main.go├── middleware├── models│   ├── models.go│   └── tag.go├── pkg│   ├── e│   │   ├── code.go│   │   └── msg.go│  ...
process
gin框架的调用流程 ray yc s blog 文件树:my ├── conf│   └── app ini├── main go├── middleware├── models│   ├── models go│   └── tag go├── pkg│   ├── e│   │   ├── code go│   │   └── msg go│  
1
779,430
27,352,489,273
IssuesEvent
2023-02-27 10:31:53
AY2223S2-CS2103T-W12-4/tp
https://api.github.com/repos/AY2223S2-CS2103T-W12-4/tp
opened
Add patient appointments
type.Task priority.High
As a doctor, I can add patient appointments so that I can schedule and keep track of all my patient visits.
1.0
Add patient appointments - As a doctor, I can add patient appointments so that I can schedule and keep track of all my patient visits.
non_process
add patient appointments as a doctor i can add patient appointments so that i can schedule and keep track of all my patient visits
0
12,160
14,741,496,131
IssuesEvent
2021-01-07 10:42:36
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
SA Billing - Winnipeg - Invalid Late Fees
anc-process anp-1 ant-bug
In GitLab by @kdjstudios on Jan 22, 2019, 16:02 **Submitted by:** ed Elizabeth Fed From: "Elizabeth Fed" <elizabeth.fed@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-01-15-44251/conversation **Server:** Internal **Client/Site:** Winnipeg **Account:** NA **Issue:** Approval to issue the credits for the late fee charges and to apply it to code 999.
1.0
SA Billing - Winnipeg - Invalid Late Fees - In GitLab by @kdjstudios on Jan 22, 2019, 16:02 **Submitted by:** ed Elizabeth Fed From: "Elizabeth Fed" <elizabeth.fed@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-01-15-44251/conversation **Server:** Internal **Client/Site:** Winnipeg **Account:** NA **Issue:** Approval to issue the credits for the late fee charges and to apply it to code 999.
process
sa billing winnipeg invalid late fees in gitlab by kdjstudios on jan submitted by ed elizabeth fed from elizabeth fed helpdesk server internal client site winnipeg account na issue approval to issue the credits for the late fee charges and to apply it to code
1
410,311
11,986,190,583
IssuesEvent
2020-04-07 18:51:21
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
opened
Disband even troops on colonization
:exclamation: priority high :question: suggestion :question:
<!-- DO NOT REMOVE PRE-EXISTING LINES IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION ---------------------------------------------------------------------------------------------------------- --> **Describe your suggestion in full detail below:** Disband event troops (try `disband_event_forces = yes`) when Jaina, Cho'gall, Durotan, etc. colonize new lands.
1.0
Disband even troops on colonization - <!-- DO NOT REMOVE PRE-EXISTING LINES IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION ---------------------------------------------------------------------------------------------------------- --> **Describe your suggestion in full detail below:** Disband event troops (try `disband_event_forces = yes`) when Jaina, Cho'gall, Durotan, etc. colonize new lands.
non_process
disband even troops on colonization do not remove pre existing lines if you want to suggest a few things open a new issue per every suggestion describe your suggestion in full detail below disband event troops try disband event forces yes when jaina cho gall durotan etc colonize new lands
0
15,139
3,927,169,028
IssuesEvent
2016-04-23 11:28:21
MarlinFirmware/Marlin
https://api.github.com/repos/MarlinFirmware/Marlin
closed
Synchronized vs unsynchronized commands
Documentation Issue Inactive
Hello! Is there any documentation about what commands are synchronized/buffered and which ones are executed immediately? I see a lot of confusion about that point. The [documentation for M400](http://www.marlinfirmware.org/index.php/M400) states: This command should rarely be needed since non-movement commands should already wait, but M400 can be useful as a workaround for badly-behaved commands. But looking at the code most of the non-movement commands such as `M104`, `M106`, `M42`, `M280` don't call `st_synchronize()` thus appear to be executed immediately (well, as soon as the command buffer is processed but not waiting the motion queue to be finished). In this GitHub issue tracker I found several conflicting statements about `M106` being synchronized and not. What's the situation? Can this be documented clearly? Thank you! :)
1.0
Synchronized vs unsynchronized commands - Hello! Is there any documentation about what commands are synchronized/buffered and which ones are executed immediately? I see a lot of confusion about that point. The [documentation for M400](http://www.marlinfirmware.org/index.php/M400) states: This command should rarely be needed since non-movement commands should already wait, but M400 can be useful as a workaround for badly-behaved commands. But looking at the code most of the non-movement commands such as `M104`, `M106`, `M42`, `M280` don't call `st_synchronize()` thus appear to be executed immediately (well, as soon as the command buffer is processed but not waiting the motion queue to be finished). In this GitHub issue tracker I found several conflicting statements about `M106` being synchronized and not. What's the situation? Can this be documented clearly? Thank you! :)
non_process
synchronized vs unsynchronized commands hello is there any documentation about what commands are synchronized buffered and which ones are executed immediately i see a lot of confusion about that point the states this command should rarely be needed since non movement commands should already wait but can be useful as a workaround for badly behaved commands but looking at the code most of the non movement commands such as don t call st synchronize thus appear to be executed immediately well as soon as the command buffer is processed but not waiting the motion queue to be finished in this github issue tracker i found several conflicting statements about being synchronized and not what s the situation can this be documented clearly thank you
0
25,915
26,098,790,711
IssuesEvent
2022-12-27 02:33:49
julianmichael/debate
https://api.github.com/repos/julianmichael/debate
closed
Make FacilitatorPanel a separate interface from DebatePage
usability
* Gets rid of the "setting up" debate state * Can allow for creating multiple debates at once or in sequence (e.g., with similar settings) * Allows us to remove the header bar in DebatePage, or at least make it more compact
True
Make FacilitatorPanel a separate interface from DebatePage - * Gets rid of the "setting up" debate state * Can allow for creating multiple debates at once or in sequence (e.g., with similar settings) * Allows us to remove the header bar in DebatePage, or at least make it more compact
non_process
make facilitatorpanel a separate interface from debatepage gets rid of the setting up debate state can allow for creating multiple debates at once or in sequence e g with similar settings allows us to remove the header bar in debatepage or at least make it more compact
0
20,795
27,541,766,372
IssuesEvent
2023-03-07 09:03:21
Fields2Cover/Fields2Cover
https://api.github.com/repos/Fields2Cover/Fields2Cover
closed
How to identify the first and last point in the swath path
processing-f2c-output
Hi @Gonzalo-Mier , Inside the swath-generated path, starting point and end point looks similar. How to identify the starting point and the endpoint.
1.0
How to identify the first and last point in the swath path - Hi @Gonzalo-Mier , Inside the swath-generated path, starting point and end point looks similar. How to identify the starting point and the endpoint.
process
how to identify the first and last point in the swath path hi gonzalo mier inside the swath generated path starting point and end point looks similar how to identify the starting point and the endpoint
1
44,365
23,598,460,472
IssuesEvent
2022-08-23 21:53:51
fastly/js-compute-runtime
https://api.github.com/repos/fastly/js-compute-runtime
opened
Copy less for request/response objects
performance
We do a lot of copying for Request and Response objects currently, as their fields may be accessed by the javascript runtime after the corresponding handles in the host have been closed. If we had the ability to clone a request or response handle with a new host call, we could keep that around instead of copying all the data across into the runtime.
True
Copy less for request/response objects - We do a lot of copying for Request and Response objects currently, as their fields may be accessed by the javascript runtime after the corresponding handles in the host have been closed. If we had the ability to clone a request or response handle with a new host call, we could keep that around instead of copying all the data across into the runtime.
non_process
copy less for request response objects we do a lot of copying for request and response objects currently as their fields may be accessed by the javascript runtime after the corresponding handles in the host have been closed if we had the ability to clone a request or response handle with a new host call we could keep that around instead of copying all the data across into the runtime
0
14,629
17,767,688,442
IssuesEvent
2021-08-30 09:38:30
googleapis/nodejs-bigtable
https://api.github.com/repos/googleapis/nodejs-bigtable
closed
Dependency Dashboard
type: process api: bigtable
This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/requestretry-6.x -->[chore(deps): update dependency requestretry to v6](../pull/933) ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/805) - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/885) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/p-queue-7.x -->[chore(deps): update dependency p-queue to v7](../pull/856) - [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/865) - [ ] <!-- recreate-branch=renovate/escape-string-regexp-5.x -->[fix(deps): update dependency escape-string-regexp to v5](../pull/861) - [ ] <!-- recreate-branch=renovate/yargs-17.x -->[fix(deps): update dependency yargs to v17](../pull/869) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/requestretry-6.x -->[chore(deps): update dependency requestretry to v6](../pull/933) ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/805) - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/885) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/p-queue-7.x -->[chore(deps): update dependency p-queue to v7](../pull/856) - [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/865) - [ ] <!-- recreate-branch=renovate/escape-string-regexp-5.x -->[fix(deps): update dependency escape-string-regexp to v5](../pull/861) - [ ] <!-- recreate-branch=renovate/yargs-17.x -->[fix(deps): update dependency yargs to v17](../pull/869) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to open these updates have all been created already click a checkbox below to force a retry rebase of any pull ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull mocha types mocha pull pull pull pull check this box to trigger a request for renovate to run again on this repository
1
334,641
29,931,039,584
IssuesEvent
2023-06-22 09:29:29
MohistMC/Mohist
https://api.github.com/repos/MohistMC/Mohist
closed
[1.19.4 because "p_150719_" is null
Wait Needs Testing
<!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** 1.19.4 **Mohist Version :** mohist-1.19.4-140-server **Operating System :** Ubuntu 22.04 **Concerned mod / plugin** : nothing **Logs :** https://github.com/Maou66/mcbug/issues/1#issue-1769172590 **Steps to Reproduce :** 1. Place an 'dispenser' on the stone.放个dispenser在石头上 2. Put water_bucket inside 'dispenser'. 把水桶放进去 3. Place a 'lever' next to 'dispenser'.旁边石头上放个把手 4. Pull the lever.拉动把手 5. Boom!!!server closed! bong!服务器炸了 **Description of issue :** Give all details about your issue
1.0
[1.19.4 because "p_150719_" is null - <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** 1.19.4 **Mohist Version :** mohist-1.19.4-140-server **Operating System :** Ubuntu 22.04 **Concerned mod / plugin** : nothing **Logs :** https://github.com/Maou66/mcbug/issues/1#issue-1769172590 **Steps to Reproduce :** 1. Place an 'dispenser' on the stone.放个dispenser在石头上 2. Put water_bucket inside 'dispenser'. 把水桶放进去 3. Place a 'lever' next to 'dispenser'.旁边石头上放个把手 4. Pull the lever.拉动把手 5. Boom!!!server closed! bong!服务器炸了 **Description of issue :** Give all details about your issue
non_process
because p is null important do not delete this line minecraft version mohist version mohist server operating system ubuntu concerned mod plugin nothing logs steps to reproduce place an dispenser on the stone 放个dispenser在石头上 put water bucket inside dispenser 把水桶放进去 place a lever next to dispenser 旁边石头上放个把手 pull the lever 拉动把手 boom!!!server closed! bong!服务器炸了 description of issue give all details about your issue
0
118,919
25,406,198,571
IssuesEvent
2022-11-22 15:27:58
WordPress/openverse
https://api.github.com/repos/WordPress/openverse
closed
Add toggleable 'redirection' mode that 301s the entire WordPress Theme
🟧 priority: high 🌟 goal: addition 💻 aspect: code
## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> As part of our iframe removal efforts, the Openverse WordPress theme located at https://github.com/WordPress/wordpress.org/tree/trunk/wordpress.org/public_html/wp-content/themes/pub/wporg-openverse needs to be updated to redirect the entire site to our new domain, openverse.wordpress.net. ## Description <!-- Describe the feature and how it solves the problem. --> We need to add a way for the theme to 301 redirect entirely to a new URL. This is expected to be openverse.wordpress.net. It might be nice to put this redirect mode behind a customizer setting, as we did for the iFrame url already in the theme. The redirects need to account for internationalization and query params need to be preserved. Some examples: 1. `https://ru.wordpress.org/openverse` should redirect to `openverse.wordpress.net/ru` 2. `https://wordpress.org/openverse/search/?q=dog` should redirect to `https://openverse.wordpress.net/search/?q=dog ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> There is no alternative. ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> ## Implementation <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in implementing this feature.
1.0
Add toggleable 'redirection' mode that 301s the entire WordPress Theme - ## Problem <!-- Describe a problem solved by this feature; or delete the section entirely. --> As part of our iframe removal efforts, the Openverse WordPress theme located at https://github.com/WordPress/wordpress.org/tree/trunk/wordpress.org/public_html/wp-content/themes/pub/wporg-openverse needs to be updated to redirect the entire site to our new domain, openverse.wordpress.net. ## Description <!-- Describe the feature and how it solves the problem. --> We need to add a way for the theme to 301 redirect entirely to a new URL. This is expected to be openverse.wordpress.net. It might be nice to put this redirect mode behind a customizer setting, as we did for the iFrame url already in the theme. The redirects need to account for internationalization and query params need to be preserved. Some examples: 1. `https://ru.wordpress.org/openverse` should redirect to `openverse.wordpress.net/ru` 2. `https://wordpress.org/openverse/search/?q=dog` should redirect to `https://openverse.wordpress.net/search/?q=dog ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> There is no alternative. ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> ## Implementation <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in implementing this feature.
non_process
add toggleable redirection mode that the entire wordpress theme problem as part of our iframe removal efforts the openverse wordpress theme located at needs to be updated to redirect the entire site to our new domain openverse wordpress net description we need to add a way for the theme to redirect entirely to a new url this is expected to be openverse wordpress net it might be nice to put this redirect mode behind a customizer setting as we did for the iframe url already in the theme the redirects need to account for internationalization and query params need to be preserved some examples should redirect to openverse wordpress net ru should redirect to alternatives there is no alternative additional context implementation 🙋 i would be interested in implementing this feature
0
63,494
26,419,899,665
IssuesEvent
2023-01-13 19:20:10
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
webapps: az webapp config set is missing support for 'virtual applications' & 'handler mapping' properties
Web Apps Service Attention App Services app-service-general
**Is your feature request related to a problem? Please describe.** This is in reference to the issue here https://github.com/Azure/azure-cli/issues/3512 **Describe the solution you'd like** Current work around has been to use az resource set, the settings is currently supported on the portal, but not Via CLI or powershell **Describe alternatives you've considered** rrent work around has been to use az resource set as seen here https://github.com/Azure/azure-cli/issues/3512 **Additional context** Add any other context or screenshots about the feature request here.
3.0
webapps: az webapp config set is missing support for 'virtual applications' & 'handler mapping' properties - **Is your feature request related to a problem? Please describe.** This is in reference to the issue here https://github.com/Azure/azure-cli/issues/3512 **Describe the solution you'd like** Current work around has been to use az resource set, the settings is currently supported on the portal, but not Via CLI or powershell **Describe alternatives you've considered** rrent work around has been to use az resource set as seen here https://github.com/Azure/azure-cli/issues/3512 **Additional context** Add any other context or screenshots about the feature request here.
non_process
webapps az webapp config set is missing support for virtual applications handler mapping properties is your feature request related to a problem please describe this is in reference to the issue here describe the solution you d like current work around has been to use az resource set the settings is currently supported on the portal but not via cli or powershell describe alternatives you ve considered rrent work around has been to use az resource set as seen here additional context add any other context or screenshots about the feature request here
0
452,210
32,052,052,533
IssuesEvent
2023-09-23 17:18:14
directus/directus
https://api.github.com/repos/directus/directus
closed
Broken link on glossary
Documentation
### Page https://docs.directus.io/user-guide/overview/glossary.html#modules ### Describe the Inaccuracy The [Settings](https://docs.directus.io/user-guide/settings/settings.html) link lead to a 404 error.
1.0
Broken link on glossary - ### Page https://docs.directus.io/user-guide/overview/glossary.html#modules ### Describe the Inaccuracy The [Settings](https://docs.directus.io/user-guide/settings/settings.html) link lead to a 404 error.
non_process
broken link on glossary page describe the inaccuracy the link lead to a error
0
19,196
25,321,324,490
IssuesEvent
2022-11-18 04:21:07
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Bazel 5.1.1 with openjdk17 server
P4 type: support / not a bug (process) team-Rules-Java
### Description of the bug: If using openjdk17 as bazel server runtime some actions may fail ``` java.lang.ExceptionInInitializerError at com.google.devtools.build.lib.actions.ParameterFile.writeContent(ParameterFile.java:118) at com.google.devtools.build.lib.actions.ParameterFile.writeParameterFile(ParameterFile.java:111) at com.google.devtools.build.lib.analysis.actions.ParameterFileWriteAction$ParamFileWriter.writeOutputFile(ParameterFileWriteAction.java:170) at com.google.devtools.build.lib.exec.FileWriteStrategy.beginWriteOutputToFile(FileWriteStrategy.java:58) at com.google.devtools.build.lib.analysis.actions.FileWriteActionContext.beginWriteOutputToFile(FileWriteActionContext.java:49) at com.google.devtools.build.lib.analysis.actions.AbstractFileWriteAction.beginExecution(AbstractFileWriteAction.java:66) at com.google.devtools.build.lib.actions.Action.execute(Action.java:133) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031) at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152) at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169) at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590) at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make java.lang.String(byte[],byte) accessible: module java.base does not "opens java.lang" to unnamed module @5e3c3cb at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:354) at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:297) at java.base/java.lang.reflect.Constructor.checkCanSetAccessible(Constructor.java:188) at java.base/java.lang.reflect.Constructor.setAccessible(Constructor.java:181) at com.google.devtools.build.lib.unsafe.StringUnsafe.<init>(StringUnsafe.java:75) at com.google.devtools.build.lib.unsafe.StringUnsafe.initInstance(StringUnsafe.java:56) at com.google.devtools.build.lib.unsafe.StringUnsafe.<clinit>(StringUnsafe.java:37) ... 21 more ``` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Haven't minimized this yet, also not yet tested with newer bazel releases ### Which operating system are you running Bazel on? NixOS ### What is the output of `bazel info release`? release 5.1.1- (@non-git) ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. NixOS nixpkgs ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? https://openjdk.org/jeps/403 since java16 default is > -illegal-access=deny disables all illegal-access operations except for those enabled by other command-line options, e.g., --add-opens. Something like `--add-opens=java.base/java.lang=ALL-UNNAMED` for server jvm flags could be a workaround somewhat related https://github.com/bazelbuild/bazel/issues/5599 reflection is still present in latest bazel source code https://github.com/bazelbuild/bazel/blob/master/src/main/java/com/google/devtools/build/lib/unsafe/StringUnsafe.java
1.0
Bazel 5.1.1 with openjdk17 server - ### Description of the bug: If using openjdk17 as bazel server runtime some actions may fail ``` java.lang.ExceptionInInitializerError at com.google.devtools.build.lib.actions.ParameterFile.writeContent(ParameterFile.java:118) at com.google.devtools.build.lib.actions.ParameterFile.writeParameterFile(ParameterFile.java:111) at com.google.devtools.build.lib.analysis.actions.ParameterFileWriteAction$ParamFileWriter.writeOutputFile(ParameterFileWriteAction.java:170) at com.google.devtools.build.lib.exec.FileWriteStrategy.beginWriteOutputToFile(FileWriteStrategy.java:58) at com.google.devtools.build.lib.analysis.actions.FileWriteActionContext.beginWriteOutputToFile(FileWriteActionContext.java:49) at com.google.devtools.build.lib.analysis.actions.AbstractFileWriteAction.beginExecution(AbstractFileWriteAction.java:66) at com.google.devtools.build.lib.actions.Action.execute(Action.java:133) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:907) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1076) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1031) at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:152) at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:91) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:492) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:856) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:349) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:169) at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:590) at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make java.lang.String(byte[],byte) accessible: module java.base does not "opens java.lang" to unnamed module @5e3c3cb at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:354) at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:297) at java.base/java.lang.reflect.Constructor.checkCanSetAccessible(Constructor.java:188) at java.base/java.lang.reflect.Constructor.setAccessible(Constructor.java:181) at com.google.devtools.build.lib.unsafe.StringUnsafe.<init>(StringUnsafe.java:75) at com.google.devtools.build.lib.unsafe.StringUnsafe.initInstance(StringUnsafe.java:56) at com.google.devtools.build.lib.unsafe.StringUnsafe.<clinit>(StringUnsafe.java:37) ... 21 more ``` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Haven't minimized this yet, also not yet tested with newer bazel releases ### Which operating system are you running Bazel on? NixOS ### What is the output of `bazel info release`? release 5.1.1- (@non-git) ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. NixOS nixpkgs ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? https://openjdk.org/jeps/403 since java16 default is > -illegal-access=deny disables all illegal-access operations except for those enabled by other command-line options, e.g., --add-opens. Something like `--add-opens=java.base/java.lang=ALL-UNNAMED` for server jvm flags could be a workaround somewhat related https://github.com/bazelbuild/bazel/issues/5599 reflection is still present in latest bazel source code https://github.com/bazelbuild/bazel/blob/master/src/main/java/com/google/devtools/build/lib/unsafe/StringUnsafe.java
process
bazel with server description of the bug if using as bazel server runtime some actions may fail java lang exceptionininitializererror at com google devtools build lib actions parameterfile writecontent parameterfile java at com google devtools build lib actions parameterfile writeparameterfile parameterfile java at com google devtools build lib analysis actions parameterfilewriteaction paramfilewriter writeoutputfile parameterfilewriteaction java at com google devtools build lib exec filewritestrategy beginwriteoutputtofile filewritestrategy java at com google devtools build lib analysis actions filewriteactioncontext beginwriteoutputtofile filewriteactioncontext java at com google devtools build lib analysis actions abstractfilewriteaction beginexecution abstractfilewriteaction java at com google devtools build lib actions action execute action java at com google devtools build lib skyframe skyframeactionexecutor execute skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner continueaction skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner run skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionstate runstatemachine actionexecutionstate java at com google devtools build lib skyframe actionexecutionstate getresultordependonfuture actionexecutionstate java at com google devtools build lib skyframe skyframeactionexecutor executeaction skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionfunction checkcacheandexecuteifneeded actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction computeinternal actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction compute actionexecutionfunction java at com google devtools build skyframe abstractparallelevaluator evaluate run abstractparallelevaluator java at com google devtools build lib concurrent abstractqueuevisitor wrappedrunnable run abstractqueuevisitor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by java lang reflect inaccessibleobjectexception unable to make java lang string byte byte accessible module java base does not opens java lang to unnamed module at java base java lang reflect accessibleobject checkcansetaccessible accessibleobject java at java base java lang reflect accessibleobject checkcansetaccessible accessibleobject java at java base java lang reflect constructor checkcansetaccessible constructor java at java base java lang reflect constructor setaccessible constructor java at com google devtools build lib unsafe stringunsafe stringunsafe java at com google devtools build lib unsafe stringunsafe initinstance stringunsafe java at com google devtools build lib unsafe stringunsafe stringunsafe java more what s the simplest easiest way to reproduce this bug please provide a minimal example if possible haven t minimized this yet also not yet tested with newer bazel releases which operating system are you running bazel on nixos what is the output of bazel info release release non git if bazel info release returns development version or non git tell us how you built bazel nixos nixpkgs what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no response any other information logs or outputs that you want to share since default is illegal access deny disables all illegal access operations except for those enabled by other command line options e g add opens something like add opens java base java lang all unnamed for server jvm flags could be a workaround somewhat related reflection is still present in latest bazel source code
1
7,410
10,532,123,464
IssuesEvent
2019-10-01 10:00:53
natario1/CameraView
https://api.github.com/repos/natario1/CameraView
closed
Camera2 Api timing issue when adding frame processors in onCameraOpened callback
about:frame processing is:bug
### Describe the bug An exception generally occurs when creating a CameraView with Camera2Engine that has a FrameProcessor adding in the CameraListener.onCameraOpended callback - CameraView version: 2.2.0 - Camera engine used: camera2 - Reproducible in official demo app: no, but probably could be - Device / Android version: Moto G7 Power Pie The exception usually occurs when the phone wakes up after sleeping and the app tries to create a new CameraView with Camera2Engine. Here is the stack trace of the error **NOTE: The error does not seem to occur when adding the FrameProcessor outside the CameraListener.onCameraOpended callback.** E/CameraEngine: uncaughtException: Unexpected exception: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.otaliastudios.cameraview.size.Size.getHeight()' on a null object reference java.lang.NullPointerException: Attempt to invoke virtual method 'int com.otaliastudios.cameraview.size.Size.getHeight()' on a null object reference at com.otaliastudios.cameraview.frame.FrameManager.setUp(FrameManager.java:116) at com.otaliastudios.cameraview.engine.Camera2Engine.onStartPreview(Camera2Engine.java:501) at com.otaliastudios.cameraview.engine.CameraEngine$10.call(CameraEngine.java:511) at com.otaliastudios.cameraview.engine.CameraEngine$10.call(CameraEngine.java:508) at com.otaliastudios.cameraview.engine.Step$2.then(Step.java:115) at com.otaliastudios.cameraview.engine.Step$2.then(Step.java:110) at com.google.android.gms.tasks.zzf.run(Unknown Source:2) at com.otaliastudios.cameraview.internal.utils.WorkerHandler.run(WorkerHandler.java:126) at com.otaliastudios.cameraview.internal.utils.WorkerHandler$1.execute(WorkerHandler.java:101) at com.google.android.gms.tasks.zze.onComplete(Unknown Source:6) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.zze(Unknown Source:132) at com.google.android.gms.tasks.zzu.continueWithTask(Unknown Source:74) at com.otaliastudios.cameraview.engine.Step.doStart(Step.java:110) at com.otaliastudios.cameraview.engine.Step.doStart(Step.java:105) at com.otaliastudios.cameraview.engine.CameraEngine.startPreview(CameraEngine.java:508) at com.otaliastudios.cameraview.engine.CameraEngine.access$400(CameraEngine.java:119) at com.otaliastudios.cameraview.engine.CameraEngine$17$1$1.then(CameraEngine.java:714) at com.otaliastudios.cameraview.engine.CameraEngine$17$1$1.then(CameraEngine.java:710) at com.google.android.gms.tasks.zzp.run(Unknown Source:2) at com.otaliastudios.cameraview.internal.utils.WorkerHandler.run(WorkerHandler.java:126) at com.otaliastudios.cameraview.internal.utils.WorkerHandler$1.execute(WorkerHandler.java:101) at com.google.android.gms.tasks.zzo.onComplete(Unknown Source:6) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.setResult(Unknown Source:86) at com.google.android.gms.tasks.zzo.onSuccess(Unknown Source:9) at com.google.android.gms.tasks.zzn.run(Unknown Source:4) at com.google.android.gms.tasks.zzt.execute(Unknown Source:2) at com.google.android.gms.tasks.zzm.onComplete(Unknown Source:11) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.setResult(Unknown Source:86) at com.google.android.gms.tasks.zzo.onSuccess(Unknown Source:9) at com.google.android.gms.tasks.zzn.run(Unknown Source:4) at com.google.android.gms.tasks.zzt.execute(Unknown Source:2) at com.google.android.gms.tasks.zzm.onComplete(Unknown Source:11) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.zze(Unknown Source:132) at com.google.android.gms.tasks.zzu.addOnSuccessListener(Unknown Source:32) at com.google.android.gms.tasks.zzp.run(Unknown Source:18) at com.otaliastudios.cameraview.internal.utils.WorkerHandler.run(WorkerHandler.java:126) at com.otaliastudios.cameraview.internal.utils.WorkerHandler$1.execute(WorkerHandler.java:101) at com.google.android.gms.tasks.zzo.onComplete(Unknown Source:6) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.setResult(Unknown Source:86) at com.google.android.gms.tasks.zze.onSuccess(Unknown Source:8) at com.google.android.gms.tasks.zzn.run(Unknown Source:4) at com.google.android.gms.tasks.zzt.execute(Unknown Source:2) at com.google.android.gms.tasks.zzm.onComplete(Unknown Source:11) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.trySetResult(Unknown Source:94) at com.google.android.gms.tasks.TaskCompletionSource.trySetResult(Unknown Source:10) at com.otaliastudios.cameraview.engine.Camera2Engine$4.onConfigured(Camera2Engine.java:472) at android.hardware.camera2.impl.CallbackProxies$SessionStateCallbackProxy.lambda$onConfigured$0(CallbackProxies.java:53) at android.hardware.camera2.impl.-$$Lambda$CallbackProxies$SessionStateCallbackProxy$soW0qC12Osypoky6AfL3P2-TeDw.run(Unknown Source:4) at android.os.Handler.handleCallback(Handler.java:873) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:193) at android.os.HandlerThread.run(HandlerThread.java:65)
1.0
Camera2 Api timing issue when adding frame processors in onCameraOpened callback - ### Describe the bug An exception generally occurs when creating a CameraView with Camera2Engine that has a FrameProcessor adding in the CameraListener.onCameraOpended callback - CameraView version: 2.2.0 - Camera engine used: camera2 - Reproducible in official demo app: no, but probably could be - Device / Android version: Moto G7 Power Pie The exception usually occurs when the phone wakes up after sleeping and the app tries to create a new CameraView with Camera2Engine. Here is the stack trace of the error **NOTE: The error does not seem to occur when adding the FrameProcessor outside the CameraListener.onCameraOpended callback.** E/CameraEngine: uncaughtException: Unexpected exception: java.lang.NullPointerException: Attempt to invoke virtual method 'int com.otaliastudios.cameraview.size.Size.getHeight()' on a null object reference java.lang.NullPointerException: Attempt to invoke virtual method 'int com.otaliastudios.cameraview.size.Size.getHeight()' on a null object reference at com.otaliastudios.cameraview.frame.FrameManager.setUp(FrameManager.java:116) at com.otaliastudios.cameraview.engine.Camera2Engine.onStartPreview(Camera2Engine.java:501) at com.otaliastudios.cameraview.engine.CameraEngine$10.call(CameraEngine.java:511) at com.otaliastudios.cameraview.engine.CameraEngine$10.call(CameraEngine.java:508) at com.otaliastudios.cameraview.engine.Step$2.then(Step.java:115) at com.otaliastudios.cameraview.engine.Step$2.then(Step.java:110) at com.google.android.gms.tasks.zzf.run(Unknown Source:2) at com.otaliastudios.cameraview.internal.utils.WorkerHandler.run(WorkerHandler.java:126) at com.otaliastudios.cameraview.internal.utils.WorkerHandler$1.execute(WorkerHandler.java:101) at com.google.android.gms.tasks.zze.onComplete(Unknown Source:6) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.zze(Unknown Source:132) at com.google.android.gms.tasks.zzu.continueWithTask(Unknown Source:74) at com.otaliastudios.cameraview.engine.Step.doStart(Step.java:110) at com.otaliastudios.cameraview.engine.Step.doStart(Step.java:105) at com.otaliastudios.cameraview.engine.CameraEngine.startPreview(CameraEngine.java:508) at com.otaliastudios.cameraview.engine.CameraEngine.access$400(CameraEngine.java:119) at com.otaliastudios.cameraview.engine.CameraEngine$17$1$1.then(CameraEngine.java:714) at com.otaliastudios.cameraview.engine.CameraEngine$17$1$1.then(CameraEngine.java:710) at com.google.android.gms.tasks.zzp.run(Unknown Source:2) at com.otaliastudios.cameraview.internal.utils.WorkerHandler.run(WorkerHandler.java:126) at com.otaliastudios.cameraview.internal.utils.WorkerHandler$1.execute(WorkerHandler.java:101) at com.google.android.gms.tasks.zzo.onComplete(Unknown Source:6) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.setResult(Unknown Source:86) at com.google.android.gms.tasks.zzo.onSuccess(Unknown Source:9) at com.google.android.gms.tasks.zzn.run(Unknown Source:4) at com.google.android.gms.tasks.zzt.execute(Unknown Source:2) at com.google.android.gms.tasks.zzm.onComplete(Unknown Source:11) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.setResult(Unknown Source:86) at com.google.android.gms.tasks.zzo.onSuccess(Unknown Source:9) at com.google.android.gms.tasks.zzn.run(Unknown Source:4) at com.google.android.gms.tasks.zzt.execute(Unknown Source:2) at com.google.android.gms.tasks.zzm.onComplete(Unknown Source:11) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.zze(Unknown Source:132) at com.google.android.gms.tasks.zzu.addOnSuccessListener(Unknown Source:32) at com.google.android.gms.tasks.zzp.run(Unknown Source:18) at com.otaliastudios.cameraview.internal.utils.WorkerHandler.run(WorkerHandler.java:126) at com.otaliastudios.cameraview.internal.utils.WorkerHandler$1.execute(WorkerHandler.java:101) at com.google.android.gms.tasks.zzo.onComplete(Unknown Source:6) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.setResult(Unknown Source:86) at com.google.android.gms.tasks.zze.onSuccess(Unknown Source:8) at com.google.android.gms.tasks.zzn.run(Unknown Source:4) at com.google.android.gms.tasks.zzt.execute(Unknown Source:2) at com.google.android.gms.tasks.zzm.onComplete(Unknown Source:11) at com.google.android.gms.tasks.zzr.zza(Unknown Source:19) at com.google.android.gms.tasks.zzu.trySetResult(Unknown Source:94) at com.google.android.gms.tasks.TaskCompletionSource.trySetResult(Unknown Source:10) at com.otaliastudios.cameraview.engine.Camera2Engine$4.onConfigured(Camera2Engine.java:472) at android.hardware.camera2.impl.CallbackProxies$SessionStateCallbackProxy.lambda$onConfigured$0(CallbackProxies.java:53) at android.hardware.camera2.impl.-$$Lambda$CallbackProxies$SessionStateCallbackProxy$soW0qC12Osypoky6AfL3P2-TeDw.run(Unknown Source:4) at android.os.Handler.handleCallback(Handler.java:873) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:193) at android.os.HandlerThread.run(HandlerThread.java:65)
process
api timing issue when adding frame processors in oncameraopened callback describe the bug an exception generally occurs when creating a cameraview with that has a frameprocessor adding in the cameralistener oncameraopended callback cameraview version camera engine used reproducible in official demo app no but probably could be device android version moto power pie the exception usually occurs when the phone wakes up after sleeping and the app tries to create a new cameraview with here is the stack trace of the error note the error does not seem to occur when adding the frameprocessor outside the cameralistener oncameraopended callback e cameraengine uncaughtexception unexpected exception java lang nullpointerexception attempt to invoke virtual method int com otaliastudios cameraview size size getheight on a null object reference java lang nullpointerexception attempt to invoke virtual method int com otaliastudios cameraview size size getheight on a null object reference at com otaliastudios cameraview frame framemanager setup framemanager java at com otaliastudios cameraview engine onstartpreview java at com otaliastudios cameraview engine cameraengine call cameraengine java at com otaliastudios cameraview engine cameraengine call cameraengine java at com otaliastudios cameraview engine step then step java at com otaliastudios cameraview engine step then step java at com google android gms tasks zzf run unknown source at com otaliastudios cameraview internal utils workerhandler run workerhandler java at com otaliastudios cameraview internal utils workerhandler execute workerhandler java at com google android gms tasks zze oncomplete unknown source at com google android gms tasks zzr zza unknown source at com google android gms tasks zzu zze unknown source at com google android gms tasks zzu continuewithtask unknown source at com otaliastudios cameraview engine step dostart step java at com otaliastudios cameraview engine step dostart step java at com otaliastudios cameraview engine cameraengine startpreview cameraengine java at com otaliastudios cameraview engine cameraengine access cameraengine java at com otaliastudios cameraview engine cameraengine then cameraengine java at com otaliastudios cameraview engine cameraengine then cameraengine java at com google android gms tasks zzp run unknown source at com otaliastudios cameraview internal utils workerhandler run workerhandler java at com otaliastudios cameraview internal utils workerhandler execute workerhandler java at com google android gms tasks zzo oncomplete unknown source at com google android gms tasks zzr zza unknown source at com google android gms tasks zzu setresult unknown source at com google android gms tasks zzo onsuccess unknown source at com google android gms tasks zzn run unknown source at com google android gms tasks zzt execute unknown source at com google android gms tasks zzm oncomplete unknown source at com google android gms tasks zzr zza unknown source at com google android gms tasks zzu setresult unknown source at com google android gms tasks zzo onsuccess unknown source at com google android gms tasks zzn run unknown source at com google android gms tasks zzt execute unknown source at com google android gms tasks zzm oncomplete unknown source at com google android gms tasks zzr zza unknown source at com google android gms tasks zzu zze unknown source at com google android gms tasks zzu addonsuccesslistener unknown source at com google android gms tasks zzp run unknown source at com otaliastudios cameraview internal utils workerhandler run workerhandler java at com otaliastudios cameraview internal utils workerhandler execute workerhandler java at com google android gms tasks zzo oncomplete unknown source at com google android gms tasks zzr zza unknown source at com google android gms tasks zzu setresult unknown source at com google android gms tasks zze onsuccess unknown source at com google android gms tasks zzn run unknown source at com google android gms tasks zzt execute unknown source at com google android gms tasks zzm oncomplete unknown source at com google android gms tasks zzr zza unknown source at com google android gms tasks zzu trysetresult unknown source at com google android gms tasks taskcompletionsource trysetresult unknown source at com otaliastudios cameraview engine onconfigured java at android hardware impl callbackproxies sessionstatecallbackproxy lambda onconfigured callbackproxies java at android hardware impl lambda callbackproxies sessionstatecallbackproxy tedw run unknown source at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android os handlerthread run handlerthread java
1
95,122
10,866,590,375
IssuesEvent
2019-11-14 21:36:41
emotion-js/emotion
https://api.github.com/repos/emotion-js/emotion
closed
what the packages
documentation
I dig emotion once it's all setup, but I'm straight up confused by the packages. I have three projects using emotion 9, via the `emotion-react` package. I'm trying to upgrade them to the latest emotion and have no fetching idea which package(s) to use now. ![image](https://user-images.githubusercontent.com/515722/65273717-081a1500-dadf-11e9-927a-528353e07893.png) This is a mess y'all. I use `css`, `injectGlobal`, `cx`, and sometimes `styled.div`. Gotta admit I'm lost with where everything went now and how to upgrade.
1.0
what the packages - I dig emotion once it's all setup, but I'm straight up confused by the packages. I have three projects using emotion 9, via the `emotion-react` package. I'm trying to upgrade them to the latest emotion and have no fetching idea which package(s) to use now. ![image](https://user-images.githubusercontent.com/515722/65273717-081a1500-dadf-11e9-927a-528353e07893.png) This is a mess y'all. I use `css`, `injectGlobal`, `cx`, and sometimes `styled.div`. Gotta admit I'm lost with where everything went now and how to upgrade.
non_process
what the packages i dig emotion once it s all setup but i m straight up confused by the packages i have three projects using emotion via the emotion react package i m trying to upgrade them to the latest emotion and have no fetching idea which package s to use now this is a mess y all i use css injectglobal cx and sometimes styled div gotta admit i m lost with where everything went now and how to upgrade
0
10,634
13,443,161,203
IssuesEvent
2020-09-08 07:56:39
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
opened
Provide limit on the `Message.Size` that the Azure Message Pump can process
area:message-processing area:security enhancement feature-request integration:service-bus
**Is your feature request related to a problem? Please describe.** Maybe we should provide the possibility to provide a maximum size limit in the Azure Service Bus message pump options that can be checked before we go through the message handling logic and maybe overload parsing/deserialization/... It can be an safety check on malicious attempts. It can then maybe be processed by a fallback message handler instead? **Describe the solution you'd like** A size limit (maybe represent metric units (KB, MB, ...) with an enumeration) in the Azure Service Bus message pump options. **Describe alternatives you've considered** Each message pump implementation and/or message handler should handle and be careful of sizes before parsing. **Additional context** When implementing the CloudEvents background job, we can use this feature to set an appropriate size limit there. See https://github.com/cloudevents/spec/blob/master/spec.md#size-limits.
1.0
Provide limit on the `Message.Size` that the Azure Message Pump can process - **Is your feature request related to a problem? Please describe.** Maybe we should provide the possibility to provide a maximum size limit in the Azure Service Bus message pump options that can be checked before we go through the message handling logic and maybe overload parsing/deserialization/... It can be an safety check on malicious attempts. It can then maybe be processed by a fallback message handler instead? **Describe the solution you'd like** A size limit (maybe represent metric units (KB, MB, ...) with an enumeration) in the Azure Service Bus message pump options. **Describe alternatives you've considered** Each message pump implementation and/or message handler should handle and be careful of sizes before parsing. **Additional context** When implementing the CloudEvents background job, we can use this feature to set an appropriate size limit there. See https://github.com/cloudevents/spec/blob/master/spec.md#size-limits.
process
provide limit on the message size that the azure message pump can process is your feature request related to a problem please describe maybe we should provide the possibility to provide a maximum size limit in the azure service bus message pump options that can be checked before we go through the message handling logic and maybe overload parsing deserialization it can be an safety check on malicious attempts it can then maybe be processed by a fallback message handler instead describe the solution you d like a size limit maybe represent metric units kb mb with an enumeration in the azure service bus message pump options describe alternatives you ve considered each message pump implementation and or message handler should handle and be careful of sizes before parsing additional context when implementing the cloudevents background job we can use this feature to set an appropriate size limit there see
1
14,154
17,065,126,708
IssuesEvent
2021-07-07 06:07:59
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Rewritten document of an iframe with the srcdoc attribute remains unproxied
AREA: client SYSTEM: iframe processing TYPE: bug
After an iframe with the `srcdoc` attribute created from script and its content dynamically written (using the document's write method), the document remains unproxied and it leads to many errors like the following: > SyntaxError: Unexpected token = in JSON at position It happens when the unproxied document tries to send a post message to the parent document. Can be reproduced with the following example: ```html <!doctype html> <html lang="en"> <head> <!-- Required meta tags --> <meta charset="utf-8"> <meta name="viewport" content="width=device-width, initial-scale=1, shrink-to-fit=no"> <title>Hello, world!</title> </head> <body> <script> var iframe = document.createElement('iframe'); iframe.srcdoc = '<h1>test</h1>'; document.body.appendChild(iframe); // iframe.contentWindow.addEventListener('load', function (e) { // console.trace('load event', e); // }); iframe.contentDocument.open(); iframe.contentDocument.write('<script>parent.postMessage("myevent", "*")<\/script>'); iframe.contentDocument.close(); </script> </body> </html> ``` Also can be reproduced on [this](https://googleads.github.io/google-publisher-tag-samples/basic/display-test-ad/demo.html) example which displays a test google ad. Based on analysis of errors which arise on the w3school.com website (separated from https://github.com/DevExpress/testcafe-hammerhead/issues/1921).
1.0
Rewritten document of an iframe with the srcdoc attribute remains unproxied - After an iframe with the `srcdoc` attribute created from script and its content dynamically written (using the document's write method), the document remains unproxied and it leads to many errors like the following: > SyntaxError: Unexpected token = in JSON at position It happens when the unproxied document tries to send a post message to the parent document. Can be reproduced with the following example: ```html <!doctype html> <html lang="en"> <head> <!-- Required meta tags --> <meta charset="utf-8"> <meta name="viewport" content="width=device-width, initial-scale=1, shrink-to-fit=no"> <title>Hello, world!</title> </head> <body> <script> var iframe = document.createElement('iframe'); iframe.srcdoc = '<h1>test</h1>'; document.body.appendChild(iframe); // iframe.contentWindow.addEventListener('load', function (e) { // console.trace('load event', e); // }); iframe.contentDocument.open(); iframe.contentDocument.write('<script>parent.postMessage("myevent", "*")<\/script>'); iframe.contentDocument.close(); </script> </body> </html> ``` Also can be reproduced on [this](https://googleads.github.io/google-publisher-tag-samples/basic/display-test-ad/demo.html) example which displays a test google ad. Based on analysis of errors which arise on the w3school.com website (separated from https://github.com/DevExpress/testcafe-hammerhead/issues/1921).
process
rewritten document of an iframe with the srcdoc attribute remains unproxied after an iframe with the srcdoc attribute created from script and its content dynamically written using the document s write method the document remains unproxied and it leads to many errors like the following syntaxerror unexpected token in json at position it happens when the unproxied document tries to send a post message to the parent document can be reproduced with the following example html hello world var iframe document createelement iframe iframe srcdoc test document body appendchild iframe iframe contentwindow addeventlistener load function e console trace load event e iframe contentdocument open iframe contentdocument write parent postmessage myevent iframe contentdocument close also can be reproduced on example which displays a test google ad based on analysis of errors which arise on the com website separated from
1
15,007
18,720,438,812
IssuesEvent
2021-11-03 11:08:38
opensafely-core/job-server
https://api.github.com/repos/opensafely-core/job-server
opened
Approving applications without `studyinformationpage` causes `RelatedObjectDoesNotExist` exception
application-process
As mentioned in #1241 and broken out into its own issue. # Reproduction 1. Start a new application. 2. Get to the contact details page and enter nothing. 3. Go to the Applications section in the Staff Area. 4. Select the application you just created. 5. Press the Approve button. # Traceback ``` RelatedObjectDoesNotExist at /staff/applications/(id)/approve/ Application has no studyinformationpage. ``` `return {"project_name": self.application.studyinformationpage.study_name}` is the cause here: ```python …/job-server/staff/views/applications.py, line 56, in get_initial return super().get_context_data(**kwargs) | { "application": self.application, } def get_initial(self): # set the value of project_name from the study_name field in the # application form return {"project_name": self.application.studyinformationpage.study_name} … @method_decorator(require_role(CoreDeveloper), name="dispatch") class ApplicationDetail(View): def get(self, request, *args, **kwargs): application = get_object_or_404( ```
1.0
Approving applications without `studyinformationpage` causes `RelatedObjectDoesNotExist` exception - As mentioned in #1241 and broken out into its own issue. # Reproduction 1. Start a new application. 2. Get to the contact details page and enter nothing. 3. Go to the Applications section in the Staff Area. 4. Select the application you just created. 5. Press the Approve button. # Traceback ``` RelatedObjectDoesNotExist at /staff/applications/(id)/approve/ Application has no studyinformationpage. ``` `return {"project_name": self.application.studyinformationpage.study_name}` is the cause here: ```python …/job-server/staff/views/applications.py, line 56, in get_initial return super().get_context_data(**kwargs) | { "application": self.application, } def get_initial(self): # set the value of project_name from the study_name field in the # application form return {"project_name": self.application.studyinformationpage.study_name} … @method_decorator(require_role(CoreDeveloper), name="dispatch") class ApplicationDetail(View): def get(self, request, *args, **kwargs): application = get_object_or_404( ```
process
approving applications without studyinformationpage causes relatedobjectdoesnotexist exception as mentioned in and broken out into its own issue reproduction start a new application get to the contact details page and enter nothing go to the applications section in the staff area select the application you just created press the approve button traceback relatedobjectdoesnotexist at staff applications id approve application has no studyinformationpage return project name self application studyinformationpage study name is the cause here python … job server staff views applications py line in get initial return super get context data kwargs application self application def get initial self set the value of project name from the study name field in the application form return project name self application studyinformationpage study name … method decorator require role coredeveloper name dispatch class applicationdetail view def get self request args kwargs application get object or
1
240,355
26,256,304,004
IssuesEvent
2023-01-06 01:14:38
attesch/PrestaShop
https://api.github.com/repos/attesch/PrestaShop
opened
WS-2020-0345 (High) detected in jsonpointer-4.0.1.tgz
security vulnerability
## WS-2020-0345 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonpointer-4.0.1.tgz</b></p></summary> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz</a></p> <p> Dependency Hierarchy: - selenium-standalone-6.14.0.tgz (Root Library) - request-2.79.0.tgz - har-validator-2.0.6.tgz - is-my-json-valid-2.17.1.tgz - :x: **jsonpointer-4.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/attesch/PrestaShop/commits/bb7465065dfeed9f49bb99b6901ab933563af739">bb7465065dfeed9f49bb99b6901ab933563af739</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype Pollution vulnerability was found in jsonpointer before 4.1.0 via the set function. <p>Publish Date: 2020-07-03 <p>URL: <a href=https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34>WS-2020-0345</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-07-03</p> <p>Fix Resolution (jsonpointer): 4.1.0</p> <p>Direct dependency fix Resolution (selenium-standalone): 6.14.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0345 (High) detected in jsonpointer-4.0.1.tgz - ## WS-2020-0345 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonpointer-4.0.1.tgz</b></p></summary> <p>Simple JSON Addressing.</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz">https://registry.npmjs.org/jsonpointer/-/jsonpointer-4.0.1.tgz</a></p> <p> Dependency Hierarchy: - selenium-standalone-6.14.0.tgz (Root Library) - request-2.79.0.tgz - har-validator-2.0.6.tgz - is-my-json-valid-2.17.1.tgz - :x: **jsonpointer-4.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/attesch/PrestaShop/commits/bb7465065dfeed9f49bb99b6901ab933563af739">bb7465065dfeed9f49bb99b6901ab933563af739</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype Pollution vulnerability was found in jsonpointer before 4.1.0 via the set function. <p>Publish Date: 2020-07-03 <p>URL: <a href=https://github.com/janl/node-jsonpointer/commit/234e3437019c6c07537ed2ad1e03b3e132b85e34>WS-2020-0345</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-07-03</p> <p>Fix Resolution (jsonpointer): 4.1.0</p> <p>Direct dependency fix Resolution (selenium-standalone): 6.14.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in jsonpointer tgz ws high severity vulnerability vulnerable library jsonpointer tgz simple json addressing library home page a href dependency hierarchy selenium standalone tgz root library request tgz har validator tgz is my json valid tgz x jsonpointer tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability was found in jsonpointer before via the set function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution jsonpointer direct dependency fix resolution selenium standalone step up your open source security game with mend
0
20,047
10,450,401,544
IssuesEvent
2019-09-19 10:27:25
symfony/symfony-docs
https://api.github.com/repos/symfony/symfony-docs
closed
[Security] Expressions: is_fully_authenticated() vs IS_AUTHENTICATED_FULLY
Security hasPR
In the [Security / Expressions](https://symfony.com/doc/4.1/security/expressions.html) docs the differences between the `is_remember_me()` expression function and the `IS_AUTHENTICATED_REMEMBERED` attribute are clearly explained in the infobox. They also mention that `is_fully_authenticated` is "Similar, but not equal to `IS_AUTHENTICATED_FULLY`, see below." but fail to provide such a distinction. Moreover, the explanation for `is_fully_authenticated` seems to be very similar the one for `IS_AUTHENTICATED_FULLY` - maybe I am missing some nuances: - [`is_fully_authenticated()`](https://symfony.com/doc/current/security/expressions.html) : > only returns true if the user has actually logged in during this session (i.e. is full-fledged) - [`IS_AUTHENTICATED_FULLY`](https://symfony.com/doc/current/security.html#checking-to-see-if-a-user-is-logged-in-is-authenticated-fully) : > Users who are logged in only because of a "remember me cookie" will have IS_AUTHENTICATED_REMEMBERED but will not have IS_AUTHENTICATED_FULLY If they are identical, we should change the explanation in the Security / Expressions docs to say : "Equal to using IS_AUTHENTICATED_FULLY with the isGranted() function." If they are not identical, we should explain how are they different. Thanks!
True
[Security] Expressions: is_fully_authenticated() vs IS_AUTHENTICATED_FULLY - In the [Security / Expressions](https://symfony.com/doc/4.1/security/expressions.html) docs the differences between the `is_remember_me()` expression function and the `IS_AUTHENTICATED_REMEMBERED` attribute are clearly explained in the infobox. They also mention that `is_fully_authenticated` is "Similar, but not equal to `IS_AUTHENTICATED_FULLY`, see below." but fail to provide such a distinction. Moreover, the explanation for `is_fully_authenticated` seems to be very similar the one for `IS_AUTHENTICATED_FULLY` - maybe I am missing some nuances: - [`is_fully_authenticated()`](https://symfony.com/doc/current/security/expressions.html) : > only returns true if the user has actually logged in during this session (i.e. is full-fledged) - [`IS_AUTHENTICATED_FULLY`](https://symfony.com/doc/current/security.html#checking-to-see-if-a-user-is-logged-in-is-authenticated-fully) : > Users who are logged in only because of a "remember me cookie" will have IS_AUTHENTICATED_REMEMBERED but will not have IS_AUTHENTICATED_FULLY If they are identical, we should change the explanation in the Security / Expressions docs to say : "Equal to using IS_AUTHENTICATED_FULLY with the isGranted() function." If they are not identical, we should explain how are they different. Thanks!
non_process
expressions is fully authenticated vs is authenticated fully in the docs the differences between the is remember me expression function and the is authenticated remembered attribute are clearly explained in the infobox they also mention that is fully authenticated is similar but not equal to is authenticated fully see below but fail to provide such a distinction moreover the explanation for is fully authenticated seems to be very similar the one for is authenticated fully maybe i am missing some nuances only returns true if the user has actually logged in during this session i e is full fledged users who are logged in only because of a remember me cookie will have is authenticated remembered but will not have is authenticated fully if they are identical we should change the explanation in the security expressions docs to say equal to using is authenticated fully with the isgranted function if they are not identical we should explain how are they different thanks
0
842
3,310,137,784
IssuesEvent
2015-11-05 06:56:11
t3kt/vjzual2
https://api.github.com/repos/t3kt/vjzual2
opened
reimplement the ramp module with more parameters
enhancement ui video processing
it should be able to cover #149 also it should be a proper reusable module
1.0
reimplement the ramp module with more parameters - it should be able to cover #149 also it should be a proper reusable module
process
reimplement the ramp module with more parameters it should be able to cover also it should be a proper reusable module
1
778,126
27,304,708,851
IssuesEvent
2023-02-24 07:02:04
FAForever/fa
https://api.github.com/repos/FAForever/fa
closed
Invincible 1HP ACUs
type: bug priority: medium
**Describe the bug** Not too sure how it happens since I've not encountered it myself, just gotten reports about it. ACU gets down to 1 hp but is unable to take any further damage and is hence invincible. **To Reproduce** ¯\_(ツ)_/¯ Here's the first time I'm aware of it happening. To Dolly around the 49 minute mark. [18846072.zip](https://github.com/FAForever/fa/files/10323740/18846072.zip) Was also recorded on a stream: https://www.twitch.tv/fartoobaked/clip/SassySavageTurnipMau5-l-rEzLt4L22Zhi_Q Also happens to Sasuke-Uchiha in this replay which isn't available yet at time of writing: [18867615](https://replay.faforever.com/18867615) **Expected behavior** [A clear and concise description of what you expected to happen.](https://www.youtube.com/watch?v=t9wmWZbr_wQ)
1.0
Invincible 1HP ACUs - **Describe the bug** Not too sure how it happens since I've not encountered it myself, just gotten reports about it. ACU gets down to 1 hp but is unable to take any further damage and is hence invincible. **To Reproduce** ¯\_(ツ)_/¯ Here's the first time I'm aware of it happening. To Dolly around the 49 minute mark. [18846072.zip](https://github.com/FAForever/fa/files/10323740/18846072.zip) Was also recorded on a stream: https://www.twitch.tv/fartoobaked/clip/SassySavageTurnipMau5-l-rEzLt4L22Zhi_Q Also happens to Sasuke-Uchiha in this replay which isn't available yet at time of writing: [18867615](https://replay.faforever.com/18867615) **Expected behavior** [A clear and concise description of what you expected to happen.](https://www.youtube.com/watch?v=t9wmWZbr_wQ)
non_process
invincible acus describe the bug not too sure how it happens since i ve not encountered it myself just gotten reports about it acu gets down to hp but is unable to take any further damage and is hence invincible to reproduce ¯ ツ ¯ here s the first time i m aware of it happening to dolly around the minute mark was also recorded on a stream also happens to sasuke uchiha in this replay which isn t available yet at time of writing expected behavior
0
768,778
26,979,839,556
IssuesEvent
2023-02-09 12:18:25
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
ETH MAC config for STM32H7X and STM32_HAL_API_V2 too late and fails
bug priority: low platform: STM32 area: Ethernet
**Describe the bug** When trying to configure an STM32H743 for 10M/half duplex it was noticed that the device was still in 100M/Full duplex. The lines changed were https://github.com/zephyrproject-rtos/zephyr/blob/main/drivers/ethernet/eth_stm32_hal.c#L1271 After adding a check of the return value of the call to `HAL_ETH_SetMACConfig()` it was noticed that it returned `HAL_ERROR`. The reason it was not noticed previously was that the return value was not checked and the device currently only runs 100M/full, which is default. The cause of the call to fail is that the device is not in `HAL_ETH_STATE_READY` as expected by `HAL_ETH_SetMACConfig()`, it is in state `HAL_ETH_STATE_BUSY` or `HAL_ETH_STATE_STARTED` (same numerical value for both states in ST HAL). By ,in a hackiish way to test, extracting out the configuration step to a separate function and then call it before `HAL_ETH_Start()` or `HAL_ETH_Start_IT()` was called the interface started to work as expected. My guess is that this bug has been around since STM32H7X support was added, since this was how it was even before `STM32_HAL_API_V2` was added. **To Reproduce** By checking the return value from `HAL_ETH_SetMACConfig()` it is easy to see that it fails. **Expected behavior** Checking for return values for potential errors. In this case, make other speeds and duplexes than 100M/Full to work. Also do configuration stage before starting Ethernet. **Impact** This is a show stopper (or at least major annoyance) for people using other than 100M/Full. I have spent weeks trying to find this. Also since we have new hardware and PHY there were many questions on where the error was. **Logs and console output** None specific. I suggest to add a check to the return value of the call to `HAL_ETH_SetMACConfig()`, and by adding a check for `!= HAL_OK` and then do a `LOG_ERR()` it should be fairly easy. **Environment (please complete the following information):** - OS: Linux - Toolchain, Zephyr SDK 0.15.1 - Zephyr Version tested 3.3.0-rc1, but bug is probably in older versions. I started in 3.2.0, where it failed as well. **Additional context** My take is that the configuration for the `STM32H7X` and `HAL_API_V2` should be a part of the configuration for the other chips. The only important part is that this configuration happens before the Ethernet is started.
1.0
ETH MAC config for STM32H7X and STM32_HAL_API_V2 too late and fails - **Describe the bug** When trying to configure an STM32H743 for 10M/half duplex it was noticed that the device was still in 100M/Full duplex. The lines changed were https://github.com/zephyrproject-rtos/zephyr/blob/main/drivers/ethernet/eth_stm32_hal.c#L1271 After adding a check of the return value of the call to `HAL_ETH_SetMACConfig()` it was noticed that it returned `HAL_ERROR`. The reason it was not noticed previously was that the return value was not checked and the device currently only runs 100M/full, which is default. The cause of the call to fail is that the device is not in `HAL_ETH_STATE_READY` as expected by `HAL_ETH_SetMACConfig()`, it is in state `HAL_ETH_STATE_BUSY` or `HAL_ETH_STATE_STARTED` (same numerical value for both states in ST HAL). By ,in a hackiish way to test, extracting out the configuration step to a separate function and then call it before `HAL_ETH_Start()` or `HAL_ETH_Start_IT()` was called the interface started to work as expected. My guess is that this bug has been around since STM32H7X support was added, since this was how it was even before `STM32_HAL_API_V2` was added. **To Reproduce** By checking the return value from `HAL_ETH_SetMACConfig()` it is easy to see that it fails. **Expected behavior** Checking for return values for potential errors. In this case, make other speeds and duplexes than 100M/Full to work. Also do configuration stage before starting Ethernet. **Impact** This is a show stopper (or at least major annoyance) for people using other than 100M/Full. I have spent weeks trying to find this. Also since we have new hardware and PHY there were many questions on where the error was. **Logs and console output** None specific. I suggest to add a check to the return value of the call to `HAL_ETH_SetMACConfig()`, and by adding a check for `!= HAL_OK` and then do a `LOG_ERR()` it should be fairly easy. **Environment (please complete the following information):** - OS: Linux - Toolchain, Zephyr SDK 0.15.1 - Zephyr Version tested 3.3.0-rc1, but bug is probably in older versions. I started in 3.2.0, where it failed as well. **Additional context** My take is that the configuration for the `STM32H7X` and `HAL_API_V2` should be a part of the configuration for the other chips. The only important part is that this configuration happens before the Ethernet is started.
non_process
eth mac config for and hal api too late and fails describe the bug when trying to configure an for half duplex it was noticed that the device was still in full duplex the lines changed were after adding a check of the return value of the call to hal eth setmacconfig it was noticed that it returned hal error the reason it was not noticed previously was that the return value was not checked and the device currently only runs full which is default the cause of the call to fail is that the device is not in hal eth state ready as expected by hal eth setmacconfig it is in state hal eth state busy or hal eth state started same numerical value for both states in st hal by in a hackiish way to test extracting out the configuration step to a separate function and then call it before hal eth start or hal eth start it was called the interface started to work as expected my guess is that this bug has been around since support was added since this was how it was even before hal api was added to reproduce by checking the return value from hal eth setmacconfig it is easy to see that it fails expected behavior checking for return values for potential errors in this case make other speeds and duplexes than full to work also do configuration stage before starting ethernet impact this is a show stopper or at least major annoyance for people using other than full i have spent weeks trying to find this also since we have new hardware and phy there were many questions on where the error was logs and console output none specific i suggest to add a check to the return value of the call to hal eth setmacconfig and by adding a check for hal ok and then do a log err it should be fairly easy environment please complete the following information os linux toolchain zephyr sdk zephyr version tested but bug is probably in older versions i started in where it failed as well additional context my take is that the configuration for the and hal api should be a part of the configuration for the other chips the only important part is that this configuration happens before the ethernet is started
0
1,802
4,540,126,524
IssuesEvent
2016-09-09 13:43:26
ongroup/mvmason
https://api.github.com/repos/ongroup/mvmason
opened
Prepare history and review of MV and SQL EUG Presentation
0 - Backlog MVON# Priority: HIGH process
<!--- @huboard:{"order":12.994801299740045,"milestone_order":0.9991004498350495} -->
1.0
Prepare history and review of MV and SQL EUG Presentation - <!--- @huboard:{"order":12.994801299740045,"milestone_order":0.9991004498350495} -->
process
prepare history and review of mv and sql eug presentation huboard order milestone order
1
5,137
7,920,765,497
IssuesEvent
2018-07-05 03:17:44
Rokid/ShadowNode
https://api.github.com/repos/Rokid/ShadowNode
closed
process: emit uncaughtException
process
the process exit directly if an uncaught exception was triggerred, should we emit an uncaughtExecption in process and not exit the process if there are any event listeners
1.0
process: emit uncaughtException - the process exit directly if an uncaught exception was triggerred, should we emit an uncaughtExecption in process and not exit the process if there are any event listeners
process
process emit uncaughtexception the process exit directly if an uncaught exception was triggerred should we emit an uncaughtexecption in process and not exit the process if there are any event listeners
1
22,564
31,789,565,337
IssuesEvent
2023-09-13 01:26:05
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
How to compile c++ JNI binary? CreateJavaVMs
P4 type: support / not a bug (process) team-Rules-Java stale
I learned to depends on `@bazel_tools//tools/jdk:jni` for JNI c++ library targets. However, I need to write a c++ program in which call Java library. How to do so? depends on `@bazel_tools//tools/jdk:jni` would generate the following error messages: ``` ld.lld: error: undefined symbol: JNI_GetCreatedJavaVMs >>> referenced by jni_helper.c >>> bazel-out/k8-fastbuild/bin/cpputil/libhdfs/_objs/hdfs/jni_helper.pic.o:(getGlobalJNIEnv) ld.lld: error: undefined symbol: JNI_CreateJavaVM >>> referenced by jni_helper.c >>> bazel-out/k8-fastbuild/bin/cpputil/libhdfs/_objs/hdfs/jni_helper.pic.o:(getGlobalJNIEnv) clang-13: error: linker command failed with exit code 1 (use -v to see invocation) ``` I also tried to put `@bazel_tools//tools/jdk:current_java_runtime` into `srcs`. But this would link against all dynamic libraries in JDK, which is not necessary. In particular, this would involve many X11 related libraries.
1.0
How to compile c++ JNI binary? CreateJavaVMs - I learned to depends on `@bazel_tools//tools/jdk:jni` for JNI c++ library targets. However, I need to write a c++ program in which call Java library. How to do so? depends on `@bazel_tools//tools/jdk:jni` would generate the following error messages: ``` ld.lld: error: undefined symbol: JNI_GetCreatedJavaVMs >>> referenced by jni_helper.c >>> bazel-out/k8-fastbuild/bin/cpputil/libhdfs/_objs/hdfs/jni_helper.pic.o:(getGlobalJNIEnv) ld.lld: error: undefined symbol: JNI_CreateJavaVM >>> referenced by jni_helper.c >>> bazel-out/k8-fastbuild/bin/cpputil/libhdfs/_objs/hdfs/jni_helper.pic.o:(getGlobalJNIEnv) clang-13: error: linker command failed with exit code 1 (use -v to see invocation) ``` I also tried to put `@bazel_tools//tools/jdk:current_java_runtime` into `srcs`. But this would link against all dynamic libraries in JDK, which is not necessary. In particular, this would involve many X11 related libraries.
process
how to compile c jni binary createjavavms i learned to depends on bazel tools tools jdk jni for jni c library targets however i need to write a c program in which call java library how to do so depends on bazel tools tools jdk jni would generate the following error messages ld lld error undefined symbol jni getcreatedjavavms referenced by jni helper c bazel out fastbuild bin cpputil libhdfs objs hdfs jni helper pic o getglobaljnienv ld lld error undefined symbol jni createjavavm referenced by jni helper c bazel out fastbuild bin cpputil libhdfs objs hdfs jni helper pic o getglobaljnienv clang error linker command failed with exit code use v to see invocation i also tried to put bazel tools tools jdk current java runtime into srcs but this would link against all dynamic libraries in jdk which is not necessary in particular this would involve many related libraries
1
8,983
12,099,948,665
IssuesEvent
2020-04-20 13:03:52
GetTerminus/terminus-ui
https://api.github.com/repos/GetTerminus/terminus-ui
closed
Demos: Should update the window title on page change
Focus: community Goal: Process Improvement Target: latest Type: chore
Currently the window title does not change when navigating around the demos. A better experience would be to update the window title on navigation. See https://angular.io/guide/set-document-title - [ ] Update the window title on navigation
1.0
Demos: Should update the window title on page change - Currently the window title does not change when navigating around the demos. A better experience would be to update the window title on navigation. See https://angular.io/guide/set-document-title - [ ] Update the window title on navigation
process
demos should update the window title on page change currently the window title does not change when navigating around the demos a better experience would be to update the window title on navigation see update the window title on navigation
1
453,025
32,079,428,424
IssuesEvent
2023-09-25 13:06:01
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
closed
Docs. Troubleshooting test run configuration with Java 17
ctg-documentation ctg-enhancement
**Description** Add Troubleshooting section in User guide on how to run tests with Java 17 using reflection. In order to support user with bug - #2026 Please consider the best place for Troubleshooting section. **Expected behavior** UnitTestBot sometimes cannot combine Object assemble model without using reflection. Sometimes a private field needs to be set or got. Reflection methods using setAccessible is forbidden in Java 17+. User can add keys to the JVm in order to run the tests. Otherwise UtUtils methods can fail. **Environment** IntelliJ IDEA Ultimate 2023.3.* **Potential alternatives** Possibility to add needed keys automatically can be very complex. One can see the list of keys needed to run tests for UnitTestBot: [build.gradle.kts](build.gradle.kts) **Context** UnitTestBot plugin will be working on Java 17 starting from Release 2023.3
1.0
Docs. Troubleshooting test run configuration with Java 17 - **Description** Add Troubleshooting section in User guide on how to run tests with Java 17 using reflection. In order to support user with bug - #2026 Please consider the best place for Troubleshooting section. **Expected behavior** UnitTestBot sometimes cannot combine Object assemble model without using reflection. Sometimes a private field needs to be set or got. Reflection methods using setAccessible is forbidden in Java 17+. User can add keys to the JVm in order to run the tests. Otherwise UtUtils methods can fail. **Environment** IntelliJ IDEA Ultimate 2023.3.* **Potential alternatives** Possibility to add needed keys automatically can be very complex. One can see the list of keys needed to run tests for UnitTestBot: [build.gradle.kts](build.gradle.kts) **Context** UnitTestBot plugin will be working on Java 17 starting from Release 2023.3
non_process
docs troubleshooting test run configuration with java description add troubleshooting section in user guide on how to run tests with java using reflection in order to support user with bug please consider the best place for troubleshooting section expected behavior unittestbot sometimes cannot combine object assemble model without using reflection sometimes a private field needs to be set or got reflection methods using setaccessible is forbidden in java user can add keys to the jvm in order to run the tests otherwise ututils methods can fail environment intellij idea ultimate potential alternatives possibility to add needed keys automatically can be very complex one can see the list of keys needed to run tests for unittestbot build gradle kts context unittestbot plugin will be working on java starting from release
0
11,452
14,272,649,016
IssuesEvent
2020-11-21 17:57:04
amor71/LiuAlgoTrader
https://api.github.com/repos/amor71/LiuAlgoTrader
closed
anchor vwap: calculation & analytics
in-process
**Is your feature request related to a problem? Please describe.** extend liu framework with anchor-vwap calculation & visualization **Describe the solution you'd like** - extend `vwap` module in `fincalcs` to support anchored vwap - revisit visualization to support anchored vwap in a notebook **Additional context** - explore better visualizations for stock charts in Jupiter Notebooks
1.0
anchor vwap: calculation & analytics - **Is your feature request related to a problem? Please describe.** extend liu framework with anchor-vwap calculation & visualization **Describe the solution you'd like** - extend `vwap` module in `fincalcs` to support anchored vwap - revisit visualization to support anchored vwap in a notebook **Additional context** - explore better visualizations for stock charts in Jupiter Notebooks
process
anchor vwap calculation analytics is your feature request related to a problem please describe extend liu framework with anchor vwap calculation visualization describe the solution you d like extend vwap module in fincalcs to support anchored vwap revisit visualization to support anchored vwap in a notebook additional context explore better visualizations for stock charts in jupiter notebooks
1
15,328
19,433,139,560
IssuesEvent
2021-12-21 14:16:01
threefoldtech/tfchain
https://api.github.com/repos/threefoldtech/tfchain
closed
POS: Alternative setup for tfchain network
process_wontfix
Currently we are planning a transition to an `npos` network instead of private network with aura. Both runtime and node code were edited in order to support babe block creation and having staking functionality. Now the following thing we want to discuss is the need for npos on devnet and testnet. Whilst it's doable to have it on devnet and testnet, we will not be able to use it in the way we want to use it for mainnet. This is because for npos we allow 3th parties to run validators and we reward them with a payout. Whilst on devnet and testnet there is no economic incentive for 3th parties to be actually running validators. Summary of networks - Devnet is a development network where TFT's on chain are worthless, meaning they are just tokens to play with and they are not pegged to real tft. - Testnet is a testing network where TFT's are pegged to real TFT's but the cost of workloads is 10% of what it is on the real (mainnet) network. - Mainnet is the real network where all things should be configured in order to be profitable for everyone involved. So why can't we run npos on devnet and testnet? ### Reasons - There is not economic incentive on devnet for 3th parties to run validators because they will have to invest in hardware in order to be able to run a validator but they won't receive any real rewards for it. - There is less economic incentive of testnet for 3th parties to run validators because the cost of workloads is 10% of what it should be on mainnet, meaning the staking pool rewards will not be filled with enought TFT's to be able to provide a good enough payout to validators. ## Solutions ### 1: Run different consensus algorithms on our networks For devnet/testnet we could stick with aura/gran consensus because there will be no need for 3th parties to run validators because they won't be rewarded good enough. #### Pro's - We don't need to migrate the current devnet & testnet. - We don't need to provide rewards for devnet & testnet. #### Con's - Alot of code changes will need to be made in order for the code to be able to support both aura/gran and babe/gran consensus. - We will need to setup another network specifically to test out the npos (babe/gran) consensus and runtime before we ever push updates to mainnet where this architecture will be used as well. - More maintenance & possible code divergence ### 2: Run npos on all networks We can run npos (babe/gran) on all the networks but the following things need to be addressed: #### Pro's - Smoother release flow (same code on all networks). - We can configure our own validators to be `invulnerables`, this will ensure us that block production and finalization will most likely not happen. - Runtime packages can be abstracted without to many changes. #### Problem 1 We need to provide a lot of staking pool rewards in order to have a good enough incentive for 3th parties to run validators on our testnet setup. And we will have to "refill" that pool if the rewards are proven to be not good enough anymore after it is running for a while. ##### Subproblem 1 Whilst having a network that is open for everyone to join and participate in consensus is good to have, there are some real caveats and reliability issues. If for example the community runs more than 60% of the validators for our chain and these nodes go down, block finalization will halt! (Block finalization will halt if not 2/3th of all validators are up and running). Making our network unrecoverable without performing a hard fork! #### Problem 2 An alternative to what is described in problem 1 is: make testnet TFT be pegged to testnet stellar TFT's. If we do this then what is actually the point of having a development network and test network whilst there could be just one test network? We could do this and also configure our own validators to be `invulnerables`, this will ensure us that block production and finalization will most likely not happen.
1.0
POS: Alternative setup for tfchain network - Currently we are planning a transition to an `npos` network instead of private network with aura. Both runtime and node code were edited in order to support babe block creation and having staking functionality. Now the following thing we want to discuss is the need for npos on devnet and testnet. Whilst it's doable to have it on devnet and testnet, we will not be able to use it in the way we want to use it for mainnet. This is because for npos we allow 3th parties to run validators and we reward them with a payout. Whilst on devnet and testnet there is no economic incentive for 3th parties to be actually running validators. Summary of networks - Devnet is a development network where TFT's on chain are worthless, meaning they are just tokens to play with and they are not pegged to real tft. - Testnet is a testing network where TFT's are pegged to real TFT's but the cost of workloads is 10% of what it is on the real (mainnet) network. - Mainnet is the real network where all things should be configured in order to be profitable for everyone involved. So why can't we run npos on devnet and testnet? ### Reasons - There is not economic incentive on devnet for 3th parties to run validators because they will have to invest in hardware in order to be able to run a validator but they won't receive any real rewards for it. - There is less economic incentive of testnet for 3th parties to run validators because the cost of workloads is 10% of what it should be on mainnet, meaning the staking pool rewards will not be filled with enought TFT's to be able to provide a good enough payout to validators. ## Solutions ### 1: Run different consensus algorithms on our networks For devnet/testnet we could stick with aura/gran consensus because there will be no need for 3th parties to run validators because they won't be rewarded good enough. #### Pro's - We don't need to migrate the current devnet & testnet. - We don't need to provide rewards for devnet & testnet. #### Con's - Alot of code changes will need to be made in order for the code to be able to support both aura/gran and babe/gran consensus. - We will need to setup another network specifically to test out the npos (babe/gran) consensus and runtime before we ever push updates to mainnet where this architecture will be used as well. - More maintenance & possible code divergence ### 2: Run npos on all networks We can run npos (babe/gran) on all the networks but the following things need to be addressed: #### Pro's - Smoother release flow (same code on all networks). - We can configure our own validators to be `invulnerables`, this will ensure us that block production and finalization will most likely not happen. - Runtime packages can be abstracted without to many changes. #### Problem 1 We need to provide a lot of staking pool rewards in order to have a good enough incentive for 3th parties to run validators on our testnet setup. And we will have to "refill" that pool if the rewards are proven to be not good enough anymore after it is running for a while. ##### Subproblem 1 Whilst having a network that is open for everyone to join and participate in consensus is good to have, there are some real caveats and reliability issues. If for example the community runs more than 60% of the validators for our chain and these nodes go down, block finalization will halt! (Block finalization will halt if not 2/3th of all validators are up and running). Making our network unrecoverable without performing a hard fork! #### Problem 2 An alternative to what is described in problem 1 is: make testnet TFT be pegged to testnet stellar TFT's. If we do this then what is actually the point of having a development network and test network whilst there could be just one test network? We could do this and also configure our own validators to be `invulnerables`, this will ensure us that block production and finalization will most likely not happen.
process
pos alternative setup for tfchain network currently we are planning a transition to an npos network instead of private network with aura both runtime and node code were edited in order to support babe block creation and having staking functionality now the following thing we want to discuss is the need for npos on devnet and testnet whilst it s doable to have it on devnet and testnet we will not be able to use it in the way we want to use it for mainnet this is because for npos we allow parties to run validators and we reward them with a payout whilst on devnet and testnet there is no economic incentive for parties to be actually running validators summary of networks devnet is a development network where tft s on chain are worthless meaning they are just tokens to play with and they are not pegged to real tft testnet is a testing network where tft s are pegged to real tft s but the cost of workloads is of what it is on the real mainnet network mainnet is the real network where all things should be configured in order to be profitable for everyone involved so why can t we run npos on devnet and testnet reasons there is not economic incentive on devnet for parties to run validators because they will have to invest in hardware in order to be able to run a validator but they won t receive any real rewards for it there is less economic incentive of testnet for parties to run validators because the cost of workloads is of what it should be on mainnet meaning the staking pool rewards will not be filled with enought tft s to be able to provide a good enough payout to validators solutions run different consensus algorithms on our networks for devnet testnet we could stick with aura gran consensus because there will be no need for parties to run validators because they won t be rewarded good enough pro s we don t need to migrate the current devnet testnet we don t need to provide rewards for devnet testnet con s alot of code changes will need to be made in order for the code to be able to support both aura gran and babe gran consensus we will need to setup another network specifically to test out the npos babe gran consensus and runtime before we ever push updates to mainnet where this architecture will be used as well more maintenance possible code divergence run npos on all networks we can run npos babe gran on all the networks but the following things need to be addressed pro s smoother release flow same code on all networks we can configure our own validators to be invulnerables this will ensure us that block production and finalization will most likely not happen runtime packages can be abstracted without to many changes problem we need to provide a lot of staking pool rewards in order to have a good enough incentive for parties to run validators on our testnet setup and we will have to refill that pool if the rewards are proven to be not good enough anymore after it is running for a while subproblem whilst having a network that is open for everyone to join and participate in consensus is good to have there are some real caveats and reliability issues if for example the community runs more than of the validators for our chain and these nodes go down block finalization will halt block finalization will halt if not of all validators are up and running making our network unrecoverable without performing a hard fork problem an alternative to what is described in problem is make testnet tft be pegged to testnet stellar tft s if we do this then what is actually the point of having a development network and test network whilst there could be just one test network we could do this and also configure our own validators to be invulnerables this will ensure us that block production and finalization will most likely not happen
1
257,805
22,217,130,844
IssuesEvent
2022-06-08 03:40:05
metaplex-foundation/metaplex
https://api.github.com/repos/metaplex-foundation/metaplex
closed
EISDIR: illegal operation on a directory, read
needs tests bug Stale
### Which package is this bug report for? candy machine cli ### Issue description EISDIR: illegal operation on a directory, read ### Command ```shell C:\Users\kunal\Desktop\moon>ts-node ~/metaplex/js/packages/cli/src/candy-machine-v2-cli.ts upload \ -e devnet \ -k ~/.config solana/devnet.json \ -cp ~\metaplex\js\packages\cli\config.json -c ~\metaplex\js\assets Error: EISDIR: illegal operation on a directory, read at Object.readSync (node:fs:727:3) at tryReadSync (node:fs:433:20) at Object.readFileSync (node:fs:479:19) at loadWalletKey (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\src\helpers\accounts.ts:526:34) at Command.<anonymous> (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts:132:40) at Command.listener [as _actionHandler] (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:481:17) at C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1237:65 at Command._chainOrCall (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1153:12) at Command._parseCommand (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1237:27) at Command._dispatchSubcommand (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1059:25) { errno: -4068, syscall: 'read', code: 'EISDIR' ``` ### Relevant log output _No response_ ### Operating system Wdow ### Priority this issue should have High (immediate attention needed) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
1.0
EISDIR: illegal operation on a directory, read - ### Which package is this bug report for? candy machine cli ### Issue description EISDIR: illegal operation on a directory, read ### Command ```shell C:\Users\kunal\Desktop\moon>ts-node ~/metaplex/js/packages/cli/src/candy-machine-v2-cli.ts upload \ -e devnet \ -k ~/.config solana/devnet.json \ -cp ~\metaplex\js\packages\cli\config.json -c ~\metaplex\js\assets Error: EISDIR: illegal operation on a directory, read at Object.readSync (node:fs:727:3) at tryReadSync (node:fs:433:20) at Object.readFileSync (node:fs:479:19) at loadWalletKey (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\src\helpers\accounts.ts:526:34) at Command.<anonymous> (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts:132:40) at Command.listener [as _actionHandler] (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:481:17) at C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1237:65 at Command._chainOrCall (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1153:12) at Command._parseCommand (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1237:27) at Command._dispatchSubcommand (C:\Users\kunal\Desktop\moon\~\metaplex\js\packages\cli\node_modules\commander\lib\command.js:1059:25) { errno: -4068, syscall: 'read', code: 'EISDIR' ``` ### Relevant log output _No response_ ### Operating system Wdow ### Priority this issue should have High (immediate attention needed) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
non_process
eisdir illegal operation on a directory read which package is this bug report for candy machine cli issue description eisdir illegal operation on a directory read command shell c users kunal desktop moon ts node metaplex js packages cli src candy machine cli ts upload e devnet k config solana devnet json cp metaplex js packages cli config json c metaplex js assets error eisdir illegal operation on a directory read at object readsync node fs at tryreadsync node fs at object readfilesync node fs at loadwalletkey c users kunal desktop moon metaplex js packages cli src helpers accounts ts at command c users kunal desktop moon metaplex js packages cli src candy machine cli ts at command listener c users kunal desktop moon metaplex js packages cli node modules commander lib command js at c users kunal desktop moon metaplex js packages cli node modules commander lib command js at command chainorcall c users kunal desktop moon metaplex js packages cli node modules commander lib command js at command parsecommand c users kunal desktop moon metaplex js packages cli node modules commander lib command js at command dispatchsubcommand c users kunal desktop moon metaplex js packages cli node modules commander lib command js errno syscall read code eisdir relevant log output no response operating system wdow priority this issue should have high immediate attention needed check the docs first i have checked the docs and it didn t solve my issue
0
284,205
30,913,603,365
IssuesEvent
2023-08-05 02:22:11
Satheesh575555/linux-4.1.15_CVE-2022-45934
https://api.github.com/repos/Satheesh575555/linux-4.1.15_CVE-2022-45934
reopened
CVE-2019-18198 (High) detected in linuxlinux-4.6
Mend: dependency security vulnerability
## CVE-2019-18198 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15_CVE-2022-45934/commit/7c0b143b43394df131d83e9aecb3c5518edc127a">7c0b143b43394df131d83e9aecb3c5518edc127a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before 5.3.4, a reference count usage error in the fib6_rule_suppress() function in the fib6 suppression feature of net/ipv6/fib6_rules.c, when handling the FIB_LOOKUP_NOREF flag, can be exploited by a local attacker to corrupt memory, aka CID-ca7a03c41753. <p>Publish Date: 2019-10-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18198>CVE-2019-18198</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198</a></p> <p>Release Date: 2019-10-31</p> <p>Fix Resolution: v5.4-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-18198 (High) detected in linuxlinux-4.6 - ## CVE-2019-18198 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15_CVE-2022-45934/commit/7c0b143b43394df131d83e9aecb3c5518edc127a">7c0b143b43394df131d83e9aecb3c5518edc127a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before 5.3.4, a reference count usage error in the fib6_rule_suppress() function in the fib6 suppression feature of net/ipv6/fib6_rules.c, when handling the FIB_LOOKUP_NOREF flag, can be exploited by a local attacker to corrupt memory, aka CID-ca7a03c41753. <p>Publish Date: 2019-10-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18198>CVE-2019-18198</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198</a></p> <p>Release Date: 2019-10-31</p> <p>Fix Resolution: v5.4-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net rules c net rules c vulnerability details in the linux kernel before a reference count usage error in the rule suppress function in the suppression feature of net rules c when handling the fib lookup noref flag can be exploited by a local attacker to corrupt memory aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
56,195
3,078,498,034
IssuesEvent
2015-08-21 10:38:00
hoodiehq/documentation
https://api.github.com/repos/hoodiehq/documentation
opened
No Subnav on Mobile makes docs hard to navigate
bug help wanted Priority: Medium
I tried to look up the deployment docs on my phone and I didn’t know how to navigate there from docs.hood.ie: ![img_0853](https://cloud.githubusercontent.com/assets/11321/9406510/56851da4-4801-11e5-913f-7bad8c314ba1.PNG) ![img_0854](https://cloud.githubusercontent.com/assets/11321/9406511/576f208e-4801-11e5-9662-c3494f02e9f5.PNG)
1.0
No Subnav on Mobile makes docs hard to navigate - I tried to look up the deployment docs on my phone and I didn’t know how to navigate there from docs.hood.ie: ![img_0853](https://cloud.githubusercontent.com/assets/11321/9406510/56851da4-4801-11e5-913f-7bad8c314ba1.PNG) ![img_0854](https://cloud.githubusercontent.com/assets/11321/9406511/576f208e-4801-11e5-9662-c3494f02e9f5.PNG)
non_process
no subnav on mobile makes docs hard to navigate i tried to look up the deployment docs on my phone and i didn’t know how to navigate there from docs hood ie
0
5,956
8,780,741,483
IssuesEvent
2018-12-19 18:14:48
knative/serving
https://api.github.com/repos/knative/serving
closed
Have a backup repository for our released artifacts
area/test-and-release kind/process
<!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area test-and-release Classify what kind of issue this is: /kind process /assign @jonjohnsonjr --> ## Expected Behavior We have a back up location for our Knative Serving artifacts. ## Actual Behavior We don't.
1.0
Have a backup repository for our released artifacts - <!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area test-and-release Classify what kind of issue this is: /kind process /assign @jonjohnsonjr --> ## Expected Behavior We have a back up location for our Knative Serving artifacts. ## Actual Behavior We don't.
process
have a backup repository for our released artifacts pro tip you can leave this block commented and it still works select the appropriate areas for your issue area test and release classify what kind of issue this is kind process assign jonjohnsonjr expected behavior we have a back up location for our knative serving artifacts actual behavior we don t
1
13,990
16,763,408,449
IssuesEvent
2021-06-14 05:02:55
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Models don't run from the browser interface
Bug Feedback Processing QGIS Browser Regression
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When, from the browser we do a `double click` or `right click`->`run Model` on a model, the model interface apear but after clicking the` run button` on the interface they dont run, nothing happens. If we load the model from the `Processing Toolbox` or from the browser `right click`->`edit Model`->`run ` it does work as expected. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> See above. Here is an example of one line and a model with the issue. [simple_line.zip](https://github.com/qgis/QGIS/files/6115829/simple_line.zip) **QGIS and OS versions** The model was initially created in QGIS 3.16.3 on Macos and it was working as expected, this issue only happen on version >= QGIS 3.16.4 The issue happens on: QGIS 3.16.4 - MacOS, Linux (Ubuntu 20.04), windows 10 QGIS 3.18.0 - MacOS, I did not tested on this version on Linux or Windows. <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** <!-- Add any other context about the problem here. -->
1.0
Models don't run from the browser interface - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When, from the browser we do a `double click` or `right click`->`run Model` on a model, the model interface apear but after clicking the` run button` on the interface they dont run, nothing happens. If we load the model from the `Processing Toolbox` or from the browser `right click`->`edit Model`->`run ` it does work as expected. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> See above. Here is an example of one line and a model with the issue. [simple_line.zip](https://github.com/qgis/QGIS/files/6115829/simple_line.zip) **QGIS and OS versions** The model was initially created in QGIS 3.16.3 on Macos and it was working as expected, this issue only happen on version >= QGIS 3.16.4 The issue happens on: QGIS 3.16.4 - MacOS, Linux (Ubuntu 20.04), windows 10 QGIS 3.18.0 - MacOS, I did not tested on this version on Linux or Windows. <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** <!-- Add any other context about the problem here. -->
process
models don t run from the browser interface bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug when from the browser we do a double click or right click run model on a model the model interface apear but after clicking the run button on the interface they dont run nothing happens if we load the model from the processing toolbox or from the browser right click edit model run it does work as expected how to reproduce see above here is an example of one line and a model with the issue qgis and os versions the model was initially created in qgis on macos and it was working as expected this issue only happen on version qgis the issue happens on qgis macos linux ubuntu windows qgis macos i did not tested on this version on linux or windows about click in the table ctrl a and then ctrl c finally paste here additional context
1
9,836
12,830,180,536
IssuesEvent
2020-07-07 01:20:30
aodn/imos-toolbox
https://api.github.com/repos/aodn/imos-toolbox
closed
Update aquadoppdw cutoff tilt angle
Priority-High Unit:Processing Unit:QC Unit:TimeSeries
From #635 Requirements: - [ ] Include Tilt Values for AquadoppDWs
1.0
Update aquadoppdw cutoff tilt angle - From #635 Requirements: - [ ] Include Tilt Values for AquadoppDWs
process
update aquadoppdw cutoff tilt angle from requirements include tilt values for aquadoppdws
1
353,129
25,103,105,059
IssuesEvent
2022-11-08 14:55:11
golang/go
https://api.github.com/repos/golang/go
closed
net: make LookupCNAME consistent between Unix and Windows, document
Documentation OS-Windows Proposal Proposal-Accepted Proposal-FinalCommentPeriod
LookupCNAME is pretty weird right now. Despite the name, it entirely ignores CNAME records on Unix. It launches `A` and `AAAA` record lookups to recursive resolvers and returns the first response name found in the `A` and `AAAA`, skipping over any `CNAME`. (and not even asking for a `CNAME`) But it documents that it does that... https://pkg.go.dev/net#LookupCNAME > A canonical name is the final name after following zero or more CNAME records. LookupCNAME does not return an error if host does not contain DNS "CNAME" records, as long as host resolves to address records. OTOH, on Windows, it does what you would expect from the name itself: it looks up CNAME records: ```go func (*Resolver) lookupCNAME(ctx context.Context, name string) (string, error) { // TODO(bradfitz): finish ctx plumbing. Nothing currently depends on this. acquireThread() defer releaseThread() var r *syscall.DNSRecord e := syscall.DnsQuery(name, syscall.DNS_TYPE_CNAME, 0, nil, &r, nil) ``` Here's a demo of a program behaving differently: ```go func main() { txt, err := net.LookupTXT("cname-to-txt.go4.org") log.Printf("LookupTXT = %q, %v", txt, err) cname, err := net.LookupCNAME("cname-to-txt.go4.org") log.Printf("cname = %q, %v", cname, err) } ``` On Linux/Mac: ``` 2021/12/10 21:19:45 LookupTXT = ["foo=bar"], <nil> 2021/12/10 21:19:45 cname = "", lookup cname-to-txt.go4.org: no such host ``` On Windows: ``` 2021/12/10 21:11:45 LookupTXT = ["foo=bar"], <nil> 2021/12/10 21:11:45 cname = "test-txt-record.go4.org.", <nil> ``` I like the Windows behavior better, FWIW. That's what I was looking for, but apparently it doesn't exist. Can we either: 1. add `LookupCNAMERecord` that actually looks up a CNAME record 2. redefine `LookupCNAME` to be like Windows, perhaps adding a `LookupCanonicalName` with the current weird Unix behavior of `LookupCNAME`? But at minimum: document whatever the rules are and make Unix and Windows match? At least in `Resolver.PreferGo` mode?
1.0
net: make LookupCNAME consistent between Unix and Windows, document - LookupCNAME is pretty weird right now. Despite the name, it entirely ignores CNAME records on Unix. It launches `A` and `AAAA` record lookups to recursive resolvers and returns the first response name found in the `A` and `AAAA`, skipping over any `CNAME`. (and not even asking for a `CNAME`) But it documents that it does that... https://pkg.go.dev/net#LookupCNAME > A canonical name is the final name after following zero or more CNAME records. LookupCNAME does not return an error if host does not contain DNS "CNAME" records, as long as host resolves to address records. OTOH, on Windows, it does what you would expect from the name itself: it looks up CNAME records: ```go func (*Resolver) lookupCNAME(ctx context.Context, name string) (string, error) { // TODO(bradfitz): finish ctx plumbing. Nothing currently depends on this. acquireThread() defer releaseThread() var r *syscall.DNSRecord e := syscall.DnsQuery(name, syscall.DNS_TYPE_CNAME, 0, nil, &r, nil) ``` Here's a demo of a program behaving differently: ```go func main() { txt, err := net.LookupTXT("cname-to-txt.go4.org") log.Printf("LookupTXT = %q, %v", txt, err) cname, err := net.LookupCNAME("cname-to-txt.go4.org") log.Printf("cname = %q, %v", cname, err) } ``` On Linux/Mac: ``` 2021/12/10 21:19:45 LookupTXT = ["foo=bar"], <nil> 2021/12/10 21:19:45 cname = "", lookup cname-to-txt.go4.org: no such host ``` On Windows: ``` 2021/12/10 21:11:45 LookupTXT = ["foo=bar"], <nil> 2021/12/10 21:11:45 cname = "test-txt-record.go4.org.", <nil> ``` I like the Windows behavior better, FWIW. That's what I was looking for, but apparently it doesn't exist. Can we either: 1. add `LookupCNAMERecord` that actually looks up a CNAME record 2. redefine `LookupCNAME` to be like Windows, perhaps adding a `LookupCanonicalName` with the current weird Unix behavior of `LookupCNAME`? But at minimum: document whatever the rules are and make Unix and Windows match? At least in `Resolver.PreferGo` mode?
non_process
net make lookupcname consistent between unix and windows document lookupcname is pretty weird right now despite the name it entirely ignores cname records on unix it launches a and aaaa record lookups to recursive resolvers and returns the first response name found in the a and aaaa skipping over any cname and not even asking for a cname but it documents that it does that a canonical name is the final name after following zero or more cname records lookupcname does not return an error if host does not contain dns cname records as long as host resolves to address records otoh on windows it does what you would expect from the name itself it looks up cname records go func resolver lookupcname ctx context context name string string error todo bradfitz finish ctx plumbing nothing currently depends on this acquirethread defer releasethread var r syscall dnsrecord e syscall dnsquery name syscall dns type cname nil r nil here s a demo of a program behaving differently go func main txt err net lookuptxt cname to txt org log printf lookuptxt q v txt err cname err net lookupcname cname to txt org log printf cname q v cname err on linux mac lookuptxt cname lookup cname to txt org no such host on windows lookuptxt cname test txt record org i like the windows behavior better fwiw that s what i was looking for but apparently it doesn t exist can we either add lookupcnamerecord that actually looks up a cname record redefine lookupcname to be like windows perhaps adding a lookupcanonicalname with the current weird unix behavior of lookupcname but at minimum document whatever the rules are and make unix and windows match at least in resolver prefergo mode
0
429,602
30,084,913,491
IssuesEvent
2023-06-29 07:54:29
Nalini1998/Project_Public
https://api.github.com/repos/Nalini1998/Project_Public
opened
10. Inside the if statement of our try block, we'll capture the data
documentation enhancement help wanted good first issue question
Inside the `if` statement of our `try` block, we'll capture the data that we need to populate our dropdown menu. To get the requested data, convert the `response` object to a `JSON` object, `await` the resolution of this method and save it to a variable called `jsonResponse`.
1.0
10. Inside the if statement of our try block, we'll capture the data - Inside the `if` statement of our `try` block, we'll capture the data that we need to populate our dropdown menu. To get the requested data, convert the `response` object to a `JSON` object, `await` the resolution of this method and save it to a variable called `jsonResponse`.
non_process
inside the if statement of our try block we ll capture the data inside the if statement of our try block we ll capture the data that we need to populate our dropdown menu to get the requested data convert the response object to a json object await the resolution of this method and save it to a variable called jsonresponse
0
695,805
23,872,769,763
IssuesEvent
2022-09-07 16:06:56
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
Create UI: Phone extension visual should remain "optional" until value entered
bug Priority3 ENTITY
Noticed a very minor UI bug when filing an SP registration. When entering the phone number for a proprietor, the phone extension field's optional visual look, dotted line, became solid line when any value is entered in the phone number field. See below. The extension field should keep its dotted line until the user enters a value in that field. ![image.png](https://images.zenhubusercontent.com/5e3c6715d85e5dced833aef6/d4c0b12a-bfb8-436e-ac80-6f4af298c342) ![image.png](https://images.zenhubusercontent.com/5e3c6715d85e5dced833aef6/f004742c-f5bb-4dcb-9d6a-798f6af82cc8)
1.0
Create UI: Phone extension visual should remain "optional" until value entered - Noticed a very minor UI bug when filing an SP registration. When entering the phone number for a proprietor, the phone extension field's optional visual look, dotted line, became solid line when any value is entered in the phone number field. See below. The extension field should keep its dotted line until the user enters a value in that field. ![image.png](https://images.zenhubusercontent.com/5e3c6715d85e5dced833aef6/d4c0b12a-bfb8-436e-ac80-6f4af298c342) ![image.png](https://images.zenhubusercontent.com/5e3c6715d85e5dced833aef6/f004742c-f5bb-4dcb-9d6a-798f6af82cc8)
non_process
create ui phone extension visual should remain optional until value entered noticed a very minor ui bug when filing an sp registration when entering the phone number for a proprietor the phone extension field s optional visual look dotted line became solid line when any value is entered in the phone number field see below the extension field should keep its dotted line until the user enters a value in that field
0
19,827
26,217,476,742
IssuesEvent
2023-01-04 12:14:34
firebase/firebase-cpp-sdk
https://api.github.com/repos/firebase/firebase-cpp-sdk
reopened
[C++] Nightly Integration Testing Report for Firestore
type: process nightly-testing
<hidden value="integration-test-status-comment"></hidden> ### ✅&nbsp; [build against repo] Integration test succeeded! Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Tue Jan 3 03:38 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3828524682)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Tue Jan 3 05:31 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3829304751)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against tip] Integration test succeeded! Requested by @sunmou99 on commit e61d6bb264633c720b1ce857717f4e9638f40279 Last updated: Wed Jan 4 03:43 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3837388037)**
1.0
[C++] Nightly Integration Testing Report for Firestore - <hidden value="integration-test-status-comment"></hidden> ### ✅&nbsp; [build against repo] Integration test succeeded! Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Tue Jan 3 03:38 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3828524682)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Tue Jan 3 05:31 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3829304751)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against tip] Integration test succeeded! Requested by @sunmou99 on commit e61d6bb264633c720b1ce857717f4e9638f40279 Last updated: Wed Jan 4 03:43 PST 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3837388037)**
process
nightly integration testing report for firestore ✅ nbsp integration test succeeded requested by on commit last updated tue jan pst ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated tue jan pst ✅ nbsp integration test succeeded requested by on commit last updated wed jan pst
1
606,312
18,759,452,273
IssuesEvent
2021-11-05 14:51:51
wasmerio/wasmer
https://api.github.com/repos/wasmerio/wasmer
closed
Metering does not work with LLVM backend on aarch64
🐞 bug 🕵️ needs investigation 📦 lib-compiler-llvm 📦 lib-middlewares priority-high
### Describe the bug Running with metering middleware on arm64 produces the following error: ``` Codegen("unknown ELF relocation 283") ``` ### Steps to reproduce The simplest way to reproduce is to replace `Cranelift` with `LLVM` in tests: ```diff $ git diff src/metering.rs diff --git a/lib/middlewares/src/metering.rs b/lib/middlewares/src/metering.rs index 4f968ed70..fd45faf13 100644 --- a/lib/middlewares/src/metering.rs +++ b/lib/middlewares/src/metering.rs @@ -283,7 +283,7 @@ mod tests { use super::*; use std::sync::Arc; - use wasmer::{imports, wat2wasm, CompilerConfig, Cranelift, Module, Store, JIT}; + use wasmer::{imports, wat2wasm, CompilerConfig, Cranelift, Module, Store, JIT, LLVM}; fn cost_function(operator: &Operator) -> u64 { match operator { @@ -312,7 +312,7 @@ mod tests { #[test] fn get_remaining_points_works() { let metering = Arc::new(Metering::new(10, cost_function)); - let mut compiler_config = Cranelift::default(); + let mut compiler_config = LLVM::default(); compiler_config.push_middleware(metering.clone()); let store = Store::new(&JIT::new(compiler_config).engine()); let module = Module::new(&store, bytecode()).unwrap(); @@ -357,7 +357,7 @@ mod tests { #[test] fn set_remaining_points_works() { let metering = Arc::new(Metering::new(10, cost_function)); - let mut compiler_config = Cranelift::default(); + let mut compiler_config = LLVM::default(); compiler_config.push_middleware(metering.clone()); let store = Store::new(&JIT::new(compiler_config).engine()); let module = Module::new(&store, bytecode()).unwrap(); ``` Test output is then: ``` running 2 tests test metering::tests::set_remaining_points_works ... FAILED test metering::tests::get_remaining_points_works ... FAILED failures: ---- metering::tests::set_remaining_points_works stdout ---- thread 'metering::tests::set_remaining_points_works' panicked at 'called `Result::unwrap()` on an `Err` value: Codegen("unknown ELF relocation 283")', lib/middlewares/src/metering.rs:363:54 ---- metering::tests::get_remaining_points_works stdout ---- thread 'metering::tests::get_remaining_points_works' panicked at 'called `Result::unwrap()` on an `Err` value: Codegen("unknown ELF relocation 283")', lib/middlewares/src/metering.rs:318:54 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace failures: metering::tests::get_remaining_points_works metering::tests::set_remaining_points_works test result: FAILED. 0 passed; 2 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.01s ``` Ideally we should probably run tests on all supported backends. ### Expected behavior Tests don't fail, no new errors on aarch64 appear. ### Actual behavior Tests fail, people question whether this is supposed to work to begin with.
1.0
Metering does not work with LLVM backend on aarch64 - ### Describe the bug Running with metering middleware on arm64 produces the following error: ``` Codegen("unknown ELF relocation 283") ``` ### Steps to reproduce The simplest way to reproduce is to replace `Cranelift` with `LLVM` in tests: ```diff $ git diff src/metering.rs diff --git a/lib/middlewares/src/metering.rs b/lib/middlewares/src/metering.rs index 4f968ed70..fd45faf13 100644 --- a/lib/middlewares/src/metering.rs +++ b/lib/middlewares/src/metering.rs @@ -283,7 +283,7 @@ mod tests { use super::*; use std::sync::Arc; - use wasmer::{imports, wat2wasm, CompilerConfig, Cranelift, Module, Store, JIT}; + use wasmer::{imports, wat2wasm, CompilerConfig, Cranelift, Module, Store, JIT, LLVM}; fn cost_function(operator: &Operator) -> u64 { match operator { @@ -312,7 +312,7 @@ mod tests { #[test] fn get_remaining_points_works() { let metering = Arc::new(Metering::new(10, cost_function)); - let mut compiler_config = Cranelift::default(); + let mut compiler_config = LLVM::default(); compiler_config.push_middleware(metering.clone()); let store = Store::new(&JIT::new(compiler_config).engine()); let module = Module::new(&store, bytecode()).unwrap(); @@ -357,7 +357,7 @@ mod tests { #[test] fn set_remaining_points_works() { let metering = Arc::new(Metering::new(10, cost_function)); - let mut compiler_config = Cranelift::default(); + let mut compiler_config = LLVM::default(); compiler_config.push_middleware(metering.clone()); let store = Store::new(&JIT::new(compiler_config).engine()); let module = Module::new(&store, bytecode()).unwrap(); ``` Test output is then: ``` running 2 tests test metering::tests::set_remaining_points_works ... FAILED test metering::tests::get_remaining_points_works ... FAILED failures: ---- metering::tests::set_remaining_points_works stdout ---- thread 'metering::tests::set_remaining_points_works' panicked at 'called `Result::unwrap()` on an `Err` value: Codegen("unknown ELF relocation 283")', lib/middlewares/src/metering.rs:363:54 ---- metering::tests::get_remaining_points_works stdout ---- thread 'metering::tests::get_remaining_points_works' panicked at 'called `Result::unwrap()` on an `Err` value: Codegen("unknown ELF relocation 283")', lib/middlewares/src/metering.rs:318:54 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace failures: metering::tests::get_remaining_points_works metering::tests::set_remaining_points_works test result: FAILED. 0 passed; 2 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.01s ``` Ideally we should probably run tests on all supported backends. ### Expected behavior Tests don't fail, no new errors on aarch64 appear. ### Actual behavior Tests fail, people question whether this is supposed to work to begin with.
non_process
metering does not work with llvm backend on describe the bug running with metering middleware on produces the following error codegen unknown elf relocation steps to reproduce the simplest way to reproduce is to replace cranelift with llvm in tests diff git diff src metering rs diff git a lib middlewares src metering rs b lib middlewares src metering rs index a lib middlewares src metering rs b lib middlewares src metering rs mod tests use super use std sync arc use wasmer imports compilerconfig cranelift module store jit use wasmer imports compilerconfig cranelift module store jit llvm fn cost function operator operator match operator mod tests fn get remaining points works let metering arc new metering new cost function let mut compiler config cranelift default let mut compiler config llvm default compiler config push middleware metering clone let store store new jit new compiler config engine let module module new store bytecode unwrap mod tests fn set remaining points works let metering arc new metering new cost function let mut compiler config cranelift default let mut compiler config llvm default compiler config push middleware metering clone let store store new jit new compiler config engine let module module new store bytecode unwrap test output is then running tests test metering tests set remaining points works failed test metering tests get remaining points works failed failures metering tests set remaining points works stdout thread metering tests set remaining points works panicked at called result unwrap on an err value codegen unknown elf relocation lib middlewares src metering rs metering tests get remaining points works stdout thread metering tests get remaining points works panicked at called result unwrap on an err value codegen unknown elf relocation lib middlewares src metering rs note run with rust backtrace environment variable to display a backtrace failures metering tests get remaining points works metering tests set remaining points works test result failed passed failed ignored measured filtered out finished in ideally we should probably run tests on all supported backends expected behavior tests don t fail no new errors on appear actual behavior tests fail people question whether this is supposed to work to begin with
0
5,728
8,570,060,118
IssuesEvent
2018-11-11 16:44:28
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Support oneshot signal handlers
feature request process
* **Version**: v6.11.3 * **Platform**: Ubuntu 14.04.5 LTS trusty, kernel: 3.13.0-119-generic, machine/CPU/platform: i686 i686 i686 * **Subsystem**: process I thought we had this solved already long ago in https://github.com/nodejs/node/pull/13894 but unfortunately it seems someone decided it's less important just because it got old. :-( It's bad UX to have a run-away program keep hanging even though a user tries to interrupt it with Ctrl+C multiple times, so I usually advise people to capture their signal handlers with `process.once()` instead of `process.on()`, but that's of limited use until node fully supports it. I'll post a gist link to a test case in a minute. Update: https://gist.github.com/mk-pmb/e48ded5215cf2f3fa54c5335aa68edd4
1.0
Support oneshot signal handlers - * **Version**: v6.11.3 * **Platform**: Ubuntu 14.04.5 LTS trusty, kernel: 3.13.0-119-generic, machine/CPU/platform: i686 i686 i686 * **Subsystem**: process I thought we had this solved already long ago in https://github.com/nodejs/node/pull/13894 but unfortunately it seems someone decided it's less important just because it got old. :-( It's bad UX to have a run-away program keep hanging even though a user tries to interrupt it with Ctrl+C multiple times, so I usually advise people to capture their signal handlers with `process.once()` instead of `process.on()`, but that's of limited use until node fully supports it. I'll post a gist link to a test case in a minute. Update: https://gist.github.com/mk-pmb/e48ded5215cf2f3fa54c5335aa68edd4
process
support oneshot signal handlers version platform ubuntu lts trusty kernel generic machine cpu platform subsystem process i thought we had this solved already long ago in but unfortunately it seems someone decided it s less important just because it got old it s bad ux to have a run away program keep hanging even though a user tries to interrupt it with ctrl c multiple times so i usually advise people to capture their signal handlers with process once instead of process on but that s of limited use until node fully supports it i ll post a gist link to a test case in a minute update
1
75,189
9,212,038,813
IssuesEvent
2019-03-09 20:31:02
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
closed
Review Mobile Functionality
Chore Design Priority: Medium Product Critical UI Redesign - Misc
Go through the Augur UI, testing various bits of functionality and making issues in CH for design or function problems that arise: - Viewing / Searching Markets - Market Display - Trading - Universe selection - Viewing Reporting pages - Disputing - Viewing Portfolio Pages - Claiming market proceeds - Claiming reporting proceeds ... We should wait to do this until we get through a lot of the redesign. -sharkcrayon
2.0
Review Mobile Functionality - Go through the Augur UI, testing various bits of functionality and making issues in CH for design or function problems that arise: - Viewing / Searching Markets - Market Display - Trading - Universe selection - Viewing Reporting pages - Disputing - Viewing Portfolio Pages - Claiming market proceeds - Claiming reporting proceeds ... We should wait to do this until we get through a lot of the redesign. -sharkcrayon
non_process
review mobile functionality go through the augur ui testing various bits of functionality and making issues in ch for design or function problems that arise viewing searching markets market display trading universe selection viewing reporting pages disputing viewing portfolio pages claiming market proceeds claiming reporting proceeds we should wait to do this until we get through a lot of the redesign sharkcrayon
0
5,554
8,394,511,735
IssuesEvent
2018-10-10 01:08:08
YousefAboElFoul/TrashScan
https://api.github.com/repos/YousefAboElFoul/TrashScan
closed
As a developer, I would like to design an algorithm to processes the captured video for classifying the items
High Priority Video Processing
Come up with an algorithm/process to detect an object and classify it
1.0
As a developer, I would like to design an algorithm to processes the captured video for classifying the items - Come up with an algorithm/process to detect an object and classify it
process
as a developer i would like to design an algorithm to processes the captured video for classifying the items come up with an algorithm process to detect an object and classify it
1
14,262
17,202,579,575
IssuesEvent
2021-07-17 15:06:44
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
closed
Implement biweekly releases for arrow-rs, parquet-rs
development-process
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.** Implement the process that will allow us to release to crates.io every 2 weeks as described in the [proposal](https://docs.google.com/document/d/1tMQ67iu8XyGGZuj--h9WQYB9inCk6c2sL_4xMTwENGc/edit?ts=60961758) See [official vote email chain](https://lists.apache.org/thread.html/rf13d8165728f1083a7d6c7374906c35af8d0b7ecff7fcf2021aa1621%40%3Cdev.arrow.apache.org%3E) I plan to use this ticket to track the work related to implementing this new process High level plan / updates: - [x] Created the active master branch - [x] PR to update `active_release` to to 4.1.0 version, and create release notes (https://github.com/apache/arrow-rs/pull/305) - [x] Create example release tarball and ask @andygrove if it would be uploadable to crates.io - [x] Create proposed release tarball and send to mailing list for a vote [Mailing List Thread](https://lists.apache.org/thread.html/rd9029f094ee0a75bb774b77bfea8624320c351cff3a4cca428f66813%40%3Cdev.arrow.apache.org%3E) - [x] (if passed) Promote release tarball to official build and then deploy to crates.io - [x] Document Process (https://github.com/apache/arrow-rs/pull/299) - [ ] Put automation in place to automatically cherry-pick PRs to `active_release` - [x] Remove release scripts left over from apache/arrow repo
1.0
Implement biweekly releases for arrow-rs, parquet-rs - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.** Implement the process that will allow us to release to crates.io every 2 weeks as described in the [proposal](https://docs.google.com/document/d/1tMQ67iu8XyGGZuj--h9WQYB9inCk6c2sL_4xMTwENGc/edit?ts=60961758) See [official vote email chain](https://lists.apache.org/thread.html/rf13d8165728f1083a7d6c7374906c35af8d0b7ecff7fcf2021aa1621%40%3Cdev.arrow.apache.org%3E) I plan to use this ticket to track the work related to implementing this new process High level plan / updates: - [x] Created the active master branch - [x] PR to update `active_release` to to 4.1.0 version, and create release notes (https://github.com/apache/arrow-rs/pull/305) - [x] Create example release tarball and ask @andygrove if it would be uploadable to crates.io - [x] Create proposed release tarball and send to mailing list for a vote [Mailing List Thread](https://lists.apache.org/thread.html/rd9029f094ee0a75bb774b77bfea8624320c351cff3a4cca428f66813%40%3Cdev.arrow.apache.org%3E) - [x] (if passed) Promote release tarball to official build and then deploy to crates.io - [x] Document Process (https://github.com/apache/arrow-rs/pull/299) - [ ] Put automation in place to automatically cherry-pick PRs to `active_release` - [x] Remove release scripts left over from apache/arrow repo
process
implement biweekly releases for arrow rs parquet rs is your feature request related to a problem or challenge please describe what you are trying to do implement the process that will allow us to release to crates io every weeks as described in the see i plan to use this ticket to track the work related to implementing this new process high level plan updates created the active master branch pr to update active release to to version and create release notes create example release tarball and ask andygrove if it would be uploadable to crates io create proposed release tarball and send to mailing list for a vote if passed promote release tarball to official build and then deploy to crates io document process put automation in place to automatically cherry pick prs to active release remove release scripts left over from apache arrow repo
1
56,261
6,971,583,801
IssuesEvent
2017-12-11 14:30:20
GetTerminus/terminus-ui
https://api.github.com/repos/GetTerminus/terminus-ui
opened
Collapsable sidebar / gutter
Focus: component Needs: design Needs: exploration Type: feature
We need a sidebar for actions etc. - This sidebar will originally be only for admin functionality. - The sidebar will receive all the actions that it can include. - Selecting an action will emit that action. - Sidebar should include a fuzzy finder search bar for quick command access - Autocomplete by typing, enter selects the highlighted option.
1.0
Collapsable sidebar / gutter - We need a sidebar for actions etc. - This sidebar will originally be only for admin functionality. - The sidebar will receive all the actions that it can include. - Selecting an action will emit that action. - Sidebar should include a fuzzy finder search bar for quick command access - Autocomplete by typing, enter selects the highlighted option.
non_process
collapsable sidebar gutter we need a sidebar for actions etc this sidebar will originally be only for admin functionality the sidebar will receive all the actions that it can include selecting an action will emit that action sidebar should include a fuzzy finder search bar for quick command access autocomplete by typing enter selects the highlighted option
0
278,678
21,092,376,748
IssuesEvent
2022-04-04 07:01:57
anhhanuman/cypress-articles
https://api.github.com/repos/anhhanuman/cypress-articles
opened
pass-object-to-javascript-function
documentation
## URL https://stackoverflow.com/questions/7764536/pass-object-to-javascript-function ## Problems ## Solutions ### - A ### - B
1.0
pass-object-to-javascript-function - ## URL https://stackoverflow.com/questions/7764536/pass-object-to-javascript-function ## Problems ## Solutions ### - A ### - B
non_process
pass object to javascript function url problems solutions a b
0
13,157
15,577,454,522
IssuesEvent
2021-03-17 13:33:31
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
[Process] Running PHP's internal webserver asynchronously yields no output
Bug Process Status: Needs Review
**Symfony version(s) affected**: 4.2.2 **Description** When I try to run PHP's internal webserver asynchronously using the symfony/process package no output is forwarded to registered callback-functions even though the webserver is started in the background and works properly. **How to reproduce** ```php <?php require_once __DIR__ . '/../vendor/autoload.php'; $process = new \Symfony\Component\Process\Process([ PHP_BINARY, '-S', 'localhost:8000', ]); $process->start(function() { // never called var_dump(func_get_args()); }); $process->waitUntil(function($type, $output) { // also never called var_dump(func_get_args()); return $output === "Press Ctrl-C to quit.\n"; }); // never reached as waitUntil() blocks until timeout reached // ... $process->stop(); ``` **Possible Solution** When I replace the php-command with another long-running process like `["watch", "ls"]` the script works as expected. **Additional context** ``` PHP 7.2.10-0ubuntu0.18.04.1 (cli) (built: Sep 13 2018 13:45:02) ( NTS ) Copyright (c) 1997-2018 The PHP Group Zend Engine v3.2.0, Copyright (c) 1998-2018 Zend Technologies with Zend OPcache v7.2.10-0ubuntu0.18.04.1, Copyright (c) 1999-2018, by Zend Technologies with Xdebug v2.6.0, Copyright (c) 2002-2018, by Derick Rethans ```
1.0
[Process] Running PHP's internal webserver asynchronously yields no output - **Symfony version(s) affected**: 4.2.2 **Description** When I try to run PHP's internal webserver asynchronously using the symfony/process package no output is forwarded to registered callback-functions even though the webserver is started in the background and works properly. **How to reproduce** ```php <?php require_once __DIR__ . '/../vendor/autoload.php'; $process = new \Symfony\Component\Process\Process([ PHP_BINARY, '-S', 'localhost:8000', ]); $process->start(function() { // never called var_dump(func_get_args()); }); $process->waitUntil(function($type, $output) { // also never called var_dump(func_get_args()); return $output === "Press Ctrl-C to quit.\n"; }); // never reached as waitUntil() blocks until timeout reached // ... $process->stop(); ``` **Possible Solution** When I replace the php-command with another long-running process like `["watch", "ls"]` the script works as expected. **Additional context** ``` PHP 7.2.10-0ubuntu0.18.04.1 (cli) (built: Sep 13 2018 13:45:02) ( NTS ) Copyright (c) 1997-2018 The PHP Group Zend Engine v3.2.0, Copyright (c) 1998-2018 Zend Technologies with Zend OPcache v7.2.10-0ubuntu0.18.04.1, Copyright (c) 1999-2018, by Zend Technologies with Xdebug v2.6.0, Copyright (c) 2002-2018, by Derick Rethans ```
process
running php s internal webserver asynchronously yields no output symfony version s affected description when i try to run php s internal webserver asynchronously using the symfony process package no output is forwarded to registered callback functions even though the webserver is started in the background and works properly how to reproduce php php require once dir vendor autoload php process new symfony component process process php binary s localhost process start function never called var dump func get args process waituntil function type output also never called var dump func get args return output press ctrl c to quit n never reached as waituntil blocks until timeout reached process stop possible solution when i replace the php command with another long running process like the script works as expected additional context php cli built sep nts copyright c the php group zend engine copyright c zend technologies with zend opcache copyright c by zend technologies with xdebug copyright c by derick rethans
1
22,067
30,592,344,672
IssuesEvent
2023-07-21 18:14:37
0xPolygonMiden/miden-vm
https://api.github.com/repos/0xPolygonMiden/miden-vm
closed
Execution options for processor and prover
enhancement processor
In some situations, we may want to provide additional info to the processor on how to execute a program. For example, we may want to specify the maximum number of cycles a program is allowed to consume (see #648). Or if we have some idea of how many cycles a program may execute for, we may give this hint to the processor so that it can allocated the required amount of memory upfront. We can do this by updating `processor::execute()` function to look as follows: ```Rust pub fn execute<A>( program: &Program, stack_inputs: StackInputs, advice_provider: A, options: ExecutionOptions, ) -> Result<ExecutionTrace, ExecutionError> where A: AdviceProvider, ``` Where `ExecutionOptions` could look like this: ```Rust pub struct ExecutionOptions { max_cycles: Option<u32>, expected_cycles: u32, // defaults to 64 if not specified } ``` We will also need to update the `ProofOptions` currently accepted by `prover::prove()` function like so (and maybe rename it to `ProvingOptions`): ```Rust pub struct ProofOptions { pub hash_fn: HashFunction, pub exec_options: ExecutionOptions, pub proof_options: WinterProofOptions, } ```
1.0
Execution options for processor and prover - In some situations, we may want to provide additional info to the processor on how to execute a program. For example, we may want to specify the maximum number of cycles a program is allowed to consume (see #648). Or if we have some idea of how many cycles a program may execute for, we may give this hint to the processor so that it can allocated the required amount of memory upfront. We can do this by updating `processor::execute()` function to look as follows: ```Rust pub fn execute<A>( program: &Program, stack_inputs: StackInputs, advice_provider: A, options: ExecutionOptions, ) -> Result<ExecutionTrace, ExecutionError> where A: AdviceProvider, ``` Where `ExecutionOptions` could look like this: ```Rust pub struct ExecutionOptions { max_cycles: Option<u32>, expected_cycles: u32, // defaults to 64 if not specified } ``` We will also need to update the `ProofOptions` currently accepted by `prover::prove()` function like so (and maybe rename it to `ProvingOptions`): ```Rust pub struct ProofOptions { pub hash_fn: HashFunction, pub exec_options: ExecutionOptions, pub proof_options: WinterProofOptions, } ```
process
execution options for processor and prover in some situations we may want to provide additional info to the processor on how to execute a program for example we may want to specify the maximum number of cycles a program is allowed to consume see or if we have some idea of how many cycles a program may execute for we may give this hint to the processor so that it can allocated the required amount of memory upfront we can do this by updating processor execute function to look as follows rust pub fn execute program program stack inputs stackinputs advice provider a options executionoptions result where a adviceprovider where executionoptions could look like this rust pub struct executionoptions max cycles option expected cycles defaults to if not specified we will also need to update the proofoptions currently accepted by prover prove function like so and maybe rename it to provingoptions rust pub struct proofoptions pub hash fn hashfunction pub exec options executionoptions pub proof options winterproofoptions
1
82,681
23,851,813,471
IssuesEvent
2022-09-06 18:41:31
bitcoin-s/bitcoin-s
https://api.github.com/repos/bitcoin-s/bitcoin-s
closed
slick 3.4.0-M1 dependency upgrade breaks jlink build
bug build dependencies db-commons
This is fixed in java 18, but unfortunately the tool we use to select java versions on CI isn't updated for java 18 (#4275). https://stackoverflow.com/a/70011064/967713 For some weird reason I can't build reproduce this bug locally, it does show up on CI, see #4342 https://github.com/bitcoin-s/bitcoin-s/commit/63df47e002f6cc6c18a095d51257f83d9a7ea1da
1.0
slick 3.4.0-M1 dependency upgrade breaks jlink build - This is fixed in java 18, but unfortunately the tool we use to select java versions on CI isn't updated for java 18 (#4275). https://stackoverflow.com/a/70011064/967713 For some weird reason I can't build reproduce this bug locally, it does show up on CI, see #4342 https://github.com/bitcoin-s/bitcoin-s/commit/63df47e002f6cc6c18a095d51257f83d9a7ea1da
non_process
slick dependency upgrade breaks jlink build this is fixed in java but unfortunately the tool we use to select java versions on ci isn t updated for java for some weird reason i can t build reproduce this bug locally it does show up on ci see
0
9,666
6,965,608,334
IssuesEvent
2017-12-09 08:31:16
dsherret/ts-simple-ast
https://api.github.com/repos/dsherret/ts-simple-ast
closed
Performance for creating AST is “quadratically”-decreasing with size
performance
I am using it for code generation from Protobuf-descriptions. The more I add, the slower it gets. I see a lot of allocations.
True
Performance for creating AST is “quadratically”-decreasing with size - I am using it for code generation from Protobuf-descriptions. The more I add, the slower it gets. I see a lot of allocations.
non_process
performance for creating ast is “quadratically” decreasing with size i am using it for code generation from protobuf descriptions the more i add the slower it gets i see a lot of allocations
0
14,460
17,537,440,054
IssuesEvent
2021-08-12 08:10:14
CGAL/cgal
https://api.github.com/repos/CGAL/cgal
closed
Missing functionality in CGAL to libpointmatcher interface
feature request Pkg::Point_set_processing_3 Has Pull Request
## Issue Details The libpointmatcher librairy give you the opportunity to do a point set registration by manually giving a weight to each point. To do it, you have to set a `GenericDescriptorOutlierFilter` in the ICP algorithm. >#### GenericDescriptorOutlierFilter description : >This filter weights matched points based on a 1D descriptor of either a single point cloud (either the reference or the reading). The descriptor values must be larger than zero. >- source (default: reference) - Point cloud from which the descriptor will be used: reference or reading >- descName (default: none) - Descriptor name used to weight paired points >- useSoftThreshold (default: 0) - If set to 1 (true), uses the value of the descriptor as a weight. If set to 0 (false), uses the parameter 'threshold' to set binary weights. - min: 0 - max: 1 >- useLargerThan (default: 1) - If set to 1 (true), values over the 'threshold' will have a weight of one. If set to 0 (false), values under the 'threshold' will have a weight of one. All other values will have a weight of zero. - min: 0 - max: 1 >- threshold (default: 0.1) - Value used to determine the binary weights - min: 0.0000001 - max: inf The construction of the ICP algorithm works all fine with the CGAL to libpointmatcher interface (see function to construct the ICP [here](https://github.com/CGAL/cgal/blob/master/Point_set_processing_3/include/CGAL/pointmatcher/compute_registration_transformation.h#L70)). The problem occurs when running the ICP algorithm. libpointmatcher search for the `descName` descriptor in the `source` point cloud, but it can't find it because the CGAL interface doesn't transfer this information. We can see in the [`construct_PM_cloud`](https://github.com/CGAL/cgal/blob/master/Point_set_processing_3/include/CGAL/pointmatcher/compute_registration_transformation.h#L294) function that the position of each point and its normal is transferred to the libpointmatcher point cloud but not other informations like the weight of the point that interests us here. The interface should be improved so that it is possible to provide to the `compute_registration_transformation` function an additional optional PropertyMap containing the weight of each point, and that the CGAL interface adds the descriptor with the name provided during the ICP configuration to the pointmatcher point cloud. Would it be possible to see how to make the GenericDescriptorOutlierFilter work with CGAL? I am not experienced enough to provide you with a good PullRequest adding this feature. Maybe @pomerlef can help us on this one ?
1.0
Missing functionality in CGAL to libpointmatcher interface - ## Issue Details The libpointmatcher librairy give you the opportunity to do a point set registration by manually giving a weight to each point. To do it, you have to set a `GenericDescriptorOutlierFilter` in the ICP algorithm. >#### GenericDescriptorOutlierFilter description : >This filter weights matched points based on a 1D descriptor of either a single point cloud (either the reference or the reading). The descriptor values must be larger than zero. >- source (default: reference) - Point cloud from which the descriptor will be used: reference or reading >- descName (default: none) - Descriptor name used to weight paired points >- useSoftThreshold (default: 0) - If set to 1 (true), uses the value of the descriptor as a weight. If set to 0 (false), uses the parameter 'threshold' to set binary weights. - min: 0 - max: 1 >- useLargerThan (default: 1) - If set to 1 (true), values over the 'threshold' will have a weight of one. If set to 0 (false), values under the 'threshold' will have a weight of one. All other values will have a weight of zero. - min: 0 - max: 1 >- threshold (default: 0.1) - Value used to determine the binary weights - min: 0.0000001 - max: inf The construction of the ICP algorithm works all fine with the CGAL to libpointmatcher interface (see function to construct the ICP [here](https://github.com/CGAL/cgal/blob/master/Point_set_processing_3/include/CGAL/pointmatcher/compute_registration_transformation.h#L70)). The problem occurs when running the ICP algorithm. libpointmatcher search for the `descName` descriptor in the `source` point cloud, but it can't find it because the CGAL interface doesn't transfer this information. We can see in the [`construct_PM_cloud`](https://github.com/CGAL/cgal/blob/master/Point_set_processing_3/include/CGAL/pointmatcher/compute_registration_transformation.h#L294) function that the position of each point and its normal is transferred to the libpointmatcher point cloud but not other informations like the weight of the point that interests us here. The interface should be improved so that it is possible to provide to the `compute_registration_transformation` function an additional optional PropertyMap containing the weight of each point, and that the CGAL interface adds the descriptor with the name provided during the ICP configuration to the pointmatcher point cloud. Would it be possible to see how to make the GenericDescriptorOutlierFilter work with CGAL? I am not experienced enough to provide you with a good PullRequest adding this feature. Maybe @pomerlef can help us on this one ?
process
missing functionality in cgal to libpointmatcher interface issue details the libpointmatcher librairy give you the opportunity to do a point set registration by manually giving a weight to each point to do it you have to set a genericdescriptoroutlierfilter in the icp algorithm genericdescriptoroutlierfilter description this filter weights matched points based on a descriptor of either a single point cloud either the reference or the reading the descriptor values must be larger than zero source default reference point cloud from which the descriptor will be used reference or reading descname default none descriptor name used to weight paired points usesoftthreshold default if set to true uses the value of the descriptor as a weight if set to false uses the parameter threshold to set binary weights min max uselargerthan default if set to true values over the threshold will have a weight of one if set to false values under the threshold will have a weight of one all other values will have a weight of zero min max threshold default value used to determine the binary weights min max inf the construction of the icp algorithm works all fine with the cgal to libpointmatcher interface see function to construct the icp the problem occurs when running the icp algorithm libpointmatcher search for the descname descriptor in the source point cloud but it can t find it because the cgal interface doesn t transfer this information we can see in the function that the position of each point and its normal is transferred to the libpointmatcher point cloud but not other informations like the weight of the point that interests us here the interface should be improved so that it is possible to provide to the compute registration transformation function an additional optional propertymap containing the weight of each point and that the cgal interface adds the descriptor with the name provided during the icp configuration to the pointmatcher point cloud would it be possible to see how to make the genericdescriptoroutlierfilter work with cgal i am not experienced enough to provide you with a good pullrequest adding this feature maybe pomerlef can help us on this one
1
674
3,145,593,187
IssuesEvent
2015-09-14 18:45:26
hammerlab/pileup.js
https://api.github.com/repos/hammerlab/pileup.js
closed
Update to Flow 0.14.0
process
This should be a big win because it restores support for type definitions for external libraries like Q and underscore. A quick attempt to update produced 100+ errors, so this won't be easy.
1.0
Update to Flow 0.14.0 - This should be a big win because it restores support for type definitions for external libraries like Q and underscore. A quick attempt to update produced 100+ errors, so this won't be easy.
process
update to flow this should be a big win because it restores support for type definitions for external libraries like q and underscore a quick attempt to update produced errors so this won t be easy
1
1,456
4,030,113,680
IssuesEvent
2016-05-18 13:17:52
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Memleak / 5-line memory machine
duplicate process stream tty
* **Version**: v6.2.0 * **Platform**: Windows 7 64-bit ``` var t = 0; while (true) { t++; process.stdout.write("keks given: " + Math.round(t / 1000000) + "M\r"); } ``` 5-line memory hog. If someone managed to slip this into any opensource code this could give fatal results.. :b Process.stdout.write will eat up as much ram as you've allocated to it. On some Windows machines I've actaully managed to use 15GB RAM (with an out-of-the-box installation of node, no extra allocations done). Output from the glorious attempt: ``` C:\Users\Eier\Desktop>node app.js keks given: 10M <--- Last few GCs ---> 244334 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 28.0 / 0 ms [alloca tion failure]. 244361 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 27.4 / 0 ms [alloca tion failure]. 244389 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 27.3 / 0 ms [alloca tion failure]. 244418 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 29.3 / 0 ms [alloca tion failure]. 244446 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 27.8 / 0 ms [alloca tion failure]. <--- JS stacktrace ---> Cannot get stack trace in GC. FATAL ERROR: Scavenger: semi-space copy Allocation failed - process out of memory ``` This should seriously be fixed soon. Even when writing something to the same line, it still eats up the memory pretty quickly. After writing the same line 10 million times, node crashed due to out of memory (it consumed about 1.2GB of RAM at this point).
1.0
Memleak / 5-line memory machine - * **Version**: v6.2.0 * **Platform**: Windows 7 64-bit ``` var t = 0; while (true) { t++; process.stdout.write("keks given: " + Math.round(t / 1000000) + "M\r"); } ``` 5-line memory hog. If someone managed to slip this into any opensource code this could give fatal results.. :b Process.stdout.write will eat up as much ram as you've allocated to it. On some Windows machines I've actaully managed to use 15GB RAM (with an out-of-the-box installation of node, no extra allocations done). Output from the glorious attempt: ``` C:\Users\Eier\Desktop>node app.js keks given: 10M <--- Last few GCs ---> 244334 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 28.0 / 0 ms [alloca tion failure]. 244361 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 27.4 / 0 ms [alloca tion failure]. 244389 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 27.3 / 0 ms [alloca tion failure]. 244418 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 29.3 / 0 ms [alloca tion failure]. 244446 ms: Scavenge 1393.6 (1434.8) -> 1393.6 (1434.8) MB, 27.8 / 0 ms [alloca tion failure]. <--- JS stacktrace ---> Cannot get stack trace in GC. FATAL ERROR: Scavenger: semi-space copy Allocation failed - process out of memory ``` This should seriously be fixed soon. Even when writing something to the same line, it still eats up the memory pretty quickly. After writing the same line 10 million times, node crashed due to out of memory (it consumed about 1.2GB of RAM at this point).
process
memleak line memory machine version platform windows bit var t while true t process stdout write keks given math round t m r line memory hog if someone managed to slip this into any opensource code this could give fatal results b process stdout write will eat up as much ram as you ve allocated to it on some windows machines i ve actaully managed to use ram with an out of the box installation of node no extra allocations done output from the glorious attempt c users eier desktop node app js keks given ms scavenge mb ms alloca tion failure ms scavenge mb ms alloca tion failure ms scavenge mb ms alloca tion failure ms scavenge mb ms alloca tion failure ms scavenge mb ms alloca tion failure cannot get stack trace in gc fatal error scavenger semi space copy allocation failed process out of memory this should seriously be fixed soon even when writing something to the same line it still eats up the memory pretty quickly after writing the same line million times node crashed due to out of memory it consumed about of ram at this point
1
22,300
30,854,055,368
IssuesEvent
2023-08-02 19:03:30
cohenlabUNC/clpipe
https://api.github.com/repos/cohenlabUNC/clpipe
closed
Add 3dTProject Implementation as Alternate Filtering Step
postprocess2 1.8.1 Req large
For now we don't have a scrub file available, so you'll need to alter the filtering workflow to be capable of running without a scrub file. After this is done and 3dTproject is registered as an available implementation of filtering, we'll need a test to make sure that the filtering works without scrubbing. Once scrubbing is available we'll also need to test that the step works with scrubbing and using the correct scrub values from that step. Finally, we will have to make sure that if the user decides to scrub AND TF, that the implementation is forced to the 3dTproject variant (and the user is warned of this)
1.0
Add 3dTProject Implementation as Alternate Filtering Step - For now we don't have a scrub file available, so you'll need to alter the filtering workflow to be capable of running without a scrub file. After this is done and 3dTproject is registered as an available implementation of filtering, we'll need a test to make sure that the filtering works without scrubbing. Once scrubbing is available we'll also need to test that the step works with scrubbing and using the correct scrub values from that step. Finally, we will have to make sure that if the user decides to scrub AND TF, that the implementation is forced to the 3dTproject variant (and the user is warned of this)
process
add implementation as alternate filtering step for now we don t have a scrub file available so you ll need to alter the filtering workflow to be capable of running without a scrub file after this is done and is registered as an available implementation of filtering we ll need a test to make sure that the filtering works without scrubbing once scrubbing is available we ll also need to test that the step works with scrubbing and using the correct scrub values from that step finally we will have to make sure that if the user decides to scrub and tf that the implementation is forced to the variant and the user is warned of this
1
21,607
30,013,397,767
IssuesEvent
2023-06-26 16:49:00
The-Data-Alchemists-Manipal/MindWave
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
closed
Road Lane line detection using OpenCV
gssoc23 level2 image-processing
## 💥 Proposal This project is detecting road lane using OpenCV, which is helpful to keep our vehicle on track. It would help in avoiding traffic and road accidents. Under GSSOC'23, please assign me this task.
1.0
Road Lane line detection using OpenCV - ## 💥 Proposal This project is detecting road lane using OpenCV, which is helpful to keep our vehicle on track. It would help in avoiding traffic and road accidents. Under GSSOC'23, please assign me this task.
process
road lane line detection using opencv 💥 proposal this project is detecting road lane using opencv which is helpful to keep our vehicle on track it would help in avoiding traffic and road accidents under gssoc please assign me this task
1
14,823
18,165,175,981
IssuesEvent
2021-09-27 13:58:21
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
reopened
Test TestBigQuery.test_transaction_info is flaky.
api: bigquery type: process testing
https://source.cloud.google.com/results/invocations/2ca6ab31-ba0f-4a6b-ab53-55943a5bac42/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-bigquery%2Fpresubmit%2Fsystem-3.8/log This test fails pretty regularly for me.
1.0
Test TestBigQuery.test_transaction_info is flaky. - https://source.cloud.google.com/results/invocations/2ca6ab31-ba0f-4a6b-ab53-55943a5bac42/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-bigquery%2Fpresubmit%2Fsystem-3.8/log This test fails pretty regularly for me.
process
test testbigquery test transaction info is flaky this test fails pretty regularly for me
1
165,865
20,624,207,942
IssuesEvent
2022-03-07 20:36:47
SmartBear/readyapi-swagger-assertion-plugin
https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin
opened
CVE-2022-0839 (High) detected in liquibase-core-3.3.3.jar
security vulnerability
## CVE-2022-0839 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>liquibase-core-3.3.3.jar</b></p></summary> <p>Liquibase is a tool for managing and executing database changes.</p> <p>Library home page: <a href="http://www.liquibase.org">http://www.liquibase.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/liquibase/liquibase-core/3.3.3/liquibase-core-3.3.3.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - ready-api-db-1.7.0.jar - :x: **liquibase-core-3.3.3.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Improper Restriction of XML External Entity Reference in GitHub repository liquibase/liquibase prior to 4.8.0. <p>Publish Date: 2022-03-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0839>CVE-2022-0839</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0839">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0839</a></p> <p>Release Date: 2022-03-04</p> <p>Fix Resolution: org.liquibase:liquibase-core:4.8.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.liquibase","packageName":"liquibase-core","packageVersion":"3.3.3","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;com.smartbear:ready-api-db:1.7.0;org.liquibase:liquibase-core:3.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.liquibase:liquibase-core:4.8.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-0839","vulnerabilityDetails":"Improper Restriction of XML External Entity Reference in GitHub repository liquibase/liquibase prior to 4.8.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0839","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2022-0839 (High) detected in liquibase-core-3.3.3.jar - ## CVE-2022-0839 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>liquibase-core-3.3.3.jar</b></p></summary> <p>Liquibase is a tool for managing and executing database changes.</p> <p>Library home page: <a href="http://www.liquibase.org">http://www.liquibase.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/liquibase/liquibase-core/3.3.3/liquibase-core-3.3.3.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - ready-api-db-1.7.0.jar - :x: **liquibase-core-3.3.3.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Improper Restriction of XML External Entity Reference in GitHub repository liquibase/liquibase prior to 4.8.0. <p>Publish Date: 2022-03-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0839>CVE-2022-0839</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0839">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0839</a></p> <p>Release Date: 2022-03-04</p> <p>Fix Resolution: org.liquibase:liquibase-core:4.8.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.liquibase","packageName":"liquibase-core","packageVersion":"3.3.3","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;com.smartbear:ready-api-db:1.7.0;org.liquibase:liquibase-core:3.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.liquibase:liquibase-core:4.8.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-0839","vulnerabilityDetails":"Improper Restriction of XML External Entity Reference in GitHub repository liquibase/liquibase prior to 4.8.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0839","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in liquibase core jar cve high severity vulnerability vulnerable library liquibase core jar liquibase is a tool for managing and executing database changes library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org liquibase liquibase core liquibase core jar dependency hierarchy ready api soapui pro jar root library ready api db jar x liquibase core jar vulnerable library found in base branch master vulnerability details improper restriction of xml external entity reference in github repository liquibase liquibase prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org liquibase liquibase core isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear ready api soapui pro com smartbear ready api db org liquibase liquibase core isminimumfixversionavailable true minimumfixversion org liquibase liquibase core isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails improper restriction of xml external entity reference in github repository liquibase liquibase prior to vulnerabilityurl
0
20,259
26,877,150,217
IssuesEvent
2023-02-05 06:39:51
pentium3/sys_reading
https://api.github.com/repos/pentium3/sys_reading
opened
Meces: Latency-efficient Rescaling via Prioritized State Migration for Stateful Distributed Stream Processing Systems
stream processing
https://www.usenix.org/system/files/atc22-gu-rong.pdf
1.0
Meces: Latency-efficient Rescaling via Prioritized State Migration for Stateful Distributed Stream Processing Systems - https://www.usenix.org/system/files/atc22-gu-rong.pdf
process
meces latency efficient rescaling via prioritized state migration for stateful distributed stream processing systems
1
11,554
14,435,294,893
IssuesEvent
2020-12-07 08:31:32
linuxdeepin/developer-center
https://api.github.com/repos/linuxdeepin/developer-center
closed
deepin15.11对ipv6支持的不好?可否升级软件包改进
other | delay processing place | WH suggest | functional behavior
<!--请将下方的缺陷汇报模板中的文字替换为您实际需要汇报的缺陷所对应的描述文字。--> <!--请保持一个 Issue 只专注一个缺陷,如果您有多个不同的缺陷需要汇报,请发起多个 Issue 。--> ## 缺陷描述 <!--简明清晰的描述你所需要汇报的缺陷(BUG)--> 1)我在deepin15.11发现firefox打不开某些网站,具体可以看下面这个帖子 https://bbs.deepin.org/forum.php?mod=viewthread&tid=196302&extra= 然后firefox设置中将network.dns.disableIPv6这个项设置为true,也就是关闭ipv6就好了 2)wine qq也有图片不能加载关闭系统ipv6处理修复这个问题 所以我想是不是deepin15.11中的某个包的ipv6处理有问题导致的这些问题,可否升级这个软件包修复这个问题,不要不管15.11 ## 复现步骤 <!-- 描述可以重现缺陷(BUG)的操作步骤,以便我们复现缺陷并进行修复 复现步骤为: 1. 打开 '...' 2. 点击 '....' 3. 滚动至 '....' 4. 缺陷 '...' 发生 --> ## 期望行为 <!--简明清晰的描述你所期望的正确行为--> 不应该因为软件使用ipv6导致应用软件出现问题 ## 截图 <!--如果适用于你所汇报的缺陷,可以附带截图来帮助描述你所遇到的缺陷(BUG)--> ## 复现环境: <!--部分缺陷可能需要在特定环境下才能复现,所以请尽可能详细的提供可能导致该缺陷的环境信息--> ### 发行版以及版本 <!--如. Linux Deepin 15.7--> deepin15.11 ### 相关的软件包版本 <!--如. dde-file-manager v1.7 (4.5.6.2-2)--> <!--如果你不确定是哪些包出了问题,你也可以考虑提供一个列表来描述那些你认为可能相关联的包以及它们的版本号--> ### 其他内容 <!--描述其他任何和你所要汇报的缺陷相关的内容,以便我们定位问题并进行处理。如果没有其它信息,你也可以移除这个段落,如果必要时我们会根据实际情况询问其它细节-->
1.0
deepin15.11对ipv6支持的不好?可否升级软件包改进 - <!--请将下方的缺陷汇报模板中的文字替换为您实际需要汇报的缺陷所对应的描述文字。--> <!--请保持一个 Issue 只专注一个缺陷,如果您有多个不同的缺陷需要汇报,请发起多个 Issue 。--> ## 缺陷描述 <!--简明清晰的描述你所需要汇报的缺陷(BUG)--> 1)我在deepin15.11发现firefox打不开某些网站,具体可以看下面这个帖子 https://bbs.deepin.org/forum.php?mod=viewthread&tid=196302&extra= 然后firefox设置中将network.dns.disableIPv6这个项设置为true,也就是关闭ipv6就好了 2)wine qq也有图片不能加载关闭系统ipv6处理修复这个问题 所以我想是不是deepin15.11中的某个包的ipv6处理有问题导致的这些问题,可否升级这个软件包修复这个问题,不要不管15.11 ## 复现步骤 <!-- 描述可以重现缺陷(BUG)的操作步骤,以便我们复现缺陷并进行修复 复现步骤为: 1. 打开 '...' 2. 点击 '....' 3. 滚动至 '....' 4. 缺陷 '...' 发生 --> ## 期望行为 <!--简明清晰的描述你所期望的正确行为--> 不应该因为软件使用ipv6导致应用软件出现问题 ## 截图 <!--如果适用于你所汇报的缺陷,可以附带截图来帮助描述你所遇到的缺陷(BUG)--> ## 复现环境: <!--部分缺陷可能需要在特定环境下才能复现,所以请尽可能详细的提供可能导致该缺陷的环境信息--> ### 发行版以及版本 <!--如. Linux Deepin 15.7--> deepin15.11 ### 相关的软件包版本 <!--如. dde-file-manager v1.7 (4.5.6.2-2)--> <!--如果你不确定是哪些包出了问题,你也可以考虑提供一个列表来描述那些你认为可能相关联的包以及它们的版本号--> ### 其他内容 <!--描述其他任何和你所要汇报的缺陷相关的内容,以便我们定位问题并进行处理。如果没有其它信息,你也可以移除这个段落,如果必要时我们会根据实际情况询问其它细节-->
process
?可否升级软件包改进 缺陷描述 ) ,具体可以看下面这个帖子 然后firefox设置中将network dns , )wine ,可否升级这个软件包修复这个问题, 复现步骤 描述可以重现缺陷(bug)的操作步骤,以便我们复现缺陷并进行修复 复现步骤为 打开 点击 滚动至 缺陷 发生 期望行为 截图 复现环境 发行版以及版本 相关的软件包版本 其他内容
1
197,783
22,606,153,517
IssuesEvent
2022-06-29 13:25:59
crouchr/learnage
https://api.github.com/repos/crouchr/learnage
closed
CVE-2021-45098 (High) detected in suricatasuricata-1.0.2 - autoclosed
security vulnerability
## CVE-2021-45098 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>suricatasuricata-1.0.2</b></p></summary> <p> <p>Suricata git repository maintained by the OISF</p> <p>Library home page: <a href=https://github.com/inliniac/suricata.git>https://github.com/inliniac/suricata.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/decode-tcp.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/decode-tcp.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/stream-tcp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Suricata before 6.0.4. It is possible to bypass/evade any HTTP-based signature by faking an RST TCP packet with random TCP options of the md5header from the client side. After the three-way handshake, it's possible to inject an RST ACK with a random TCP md5header option. Then, the client can send an HTTP GET request with a forbidden URL. The server will ignore the RST ACK and send the response HTTP packet for the client's request. These packets will not trigger a Suricata reject action. <p>Publish Date: 2021-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-45098>CVE-2021-45098</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-45098">https://nvd.nist.gov/vuln/detail/CVE-2021-45098</a></p> <p>Release Date: 2021-12-16</p> <p>Fix Resolution: suricata - 1:6.0.4-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-45098 (High) detected in suricatasuricata-1.0.2 - autoclosed - ## CVE-2021-45098 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>suricatasuricata-1.0.2</b></p></summary> <p> <p>Suricata git repository maintained by the OISF</p> <p>Library home page: <a href=https://github.com/inliniac/suricata.git>https://github.com/inliniac/suricata.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/decode-tcp.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/decode-tcp.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/blackrain2020/original-sources-3rd-party/suricata-1.0.2.tar/suricata-1.0.2/src/stream-tcp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Suricata before 6.0.4. It is possible to bypass/evade any HTTP-based signature by faking an RST TCP packet with random TCP options of the md5header from the client side. After the three-way handshake, it's possible to inject an RST ACK with a random TCP md5header option. Then, the client can send an HTTP GET request with a forbidden URL. The server will ignore the RST ACK and send the response HTTP packet for the client's request. These packets will not trigger a Suricata reject action. <p>Publish Date: 2021-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-45098>CVE-2021-45098</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-45098">https://nvd.nist.gov/vuln/detail/CVE-2021-45098</a></p> <p>Release Date: 2021-12-16</p> <p>Fix Resolution: suricata - 1:6.0.4-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in suricatasuricata autoclosed cve high severity vulnerability vulnerable library suricatasuricata suricata git repository maintained by the oisf library home page a href found in base branch master vulnerable source files original sources party suricata tar suricata src decode tcp c original sources party suricata tar suricata src decode tcp h original sources party suricata tar suricata src stream tcp c vulnerability details an issue was discovered in suricata before it is possible to bypass evade any http based signature by faking an rst tcp packet with random tcp options of the from the client side after the three way handshake it s possible to inject an rst ack with a random tcp option then the client can send an http get request with a forbidden url the server will ignore the rst ack and send the response http packet for the client s request these packets will not trigger a suricata reject action publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution suricata step up your open source security game with mend
0
1,238
3,777,095,587
IssuesEvent
2016-03-17 18:48:05
mapbox/mapbox-gl-js
https://api.github.com/repos/mapbox/mapbox-gl-js
opened
Move buffer stats to a separate debug page
meta ready testing & release process
![screen shot 2016-03-17 at 11 44 28 am](https://cloud.githubusercontent.com/assets/281306/13857120/b94566fa-ec35-11e5-9fb9-f0d06ab3cb5d.png) I want to keep the `index.html` debug page simple. "debug" and "collision debug" modes can stay.
1.0
Move buffer stats to a separate debug page - ![screen shot 2016-03-17 at 11 44 28 am](https://cloud.githubusercontent.com/assets/281306/13857120/b94566fa-ec35-11e5-9fb9-f0d06ab3cb5d.png) I want to keep the `index.html` debug page simple. "debug" and "collision debug" modes can stay.
process
move buffer stats to a separate debug page i want to keep the index html debug page simple debug and collision debug modes can stay
1
443,072
12,759,144,501
IssuesEvent
2020-06-29 04:55:23
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Add cookie policy and password policy text customizations
Affected/3.1.0 Priority/Normal Type/Bug WUM
### Description: Along with changes made in https://github.com/wso2/product-apim/commit/da729a0bce3c93b0fb339a579fa2b9bf40869e33#diff-22af8a23b712342a08c30c2e4b22dbb4 the cookie policy and password policy text customizability added with https://github.com/wso2/product-apim/commit/f0d5e260d863c9105224926c3e51f11200c2446f#diff-22af8a23b712342a08c30c2e4b22dbb4 has been reverted. We need to add it back. ### Affected Product Version: APIM 3.1.0
1.0
Add cookie policy and password policy text customizations - ### Description: Along with changes made in https://github.com/wso2/product-apim/commit/da729a0bce3c93b0fb339a579fa2b9bf40869e33#diff-22af8a23b712342a08c30c2e4b22dbb4 the cookie policy and password policy text customizability added with https://github.com/wso2/product-apim/commit/f0d5e260d863c9105224926c3e51f11200c2446f#diff-22af8a23b712342a08c30c2e4b22dbb4 has been reverted. We need to add it back. ### Affected Product Version: APIM 3.1.0
non_process
add cookie policy and password policy text customizations description along with changes made in the cookie policy and password policy text customizability added with has been reverted we need to add it back affected product version apim
0
7,817
5,224,809,248
IssuesEvent
2017-01-27 16:26:45
MISP/MISP
https://api.github.com/repos/MISP/MISP
closed
Upload attachement button missing
bug usability visual
Within an event when clicking on add attribute icon (+) the opened pane does not include the upload attachment button as indicated in the tooltip. Screenshot attached. ![upload](https://cloud.githubusercontent.com/assets/414086/22368561/737b4d2e-e487-11e6-85af-d4e6c04d7bba.jpg)
True
Upload attachement button missing - Within an event when clicking on add attribute icon (+) the opened pane does not include the upload attachment button as indicated in the tooltip. Screenshot attached. ![upload](https://cloud.githubusercontent.com/assets/414086/22368561/737b4d2e-e487-11e6-85af-d4e6c04d7bba.jpg)
non_process
upload attachement button missing within an event when clicking on add attribute icon the opened pane does not include the upload attachment button as indicated in the tooltip screenshot attached
0
34,073
2,775,047,941
IssuesEvent
2015-05-04 13:56:05
anchorcms/anchor-cms
https://api.github.com/repos/anchorcms/anchor-cms
closed
Deleting a page that is assigned as Home Page makes the CMS unusable
Bug Confirmed Low Priority
I deleted a page (the default posts page) that was assigned as Home Page and after that I could not use the admin nor the page itself anymore because of: ``` Uncaught Exception Trying to get property of non-object Origin anchor/views/partials/header.php on line 47 Trace #0 path/anchor/anchor/views/partials/header.php(47): System\Error::native(8, 'Trying to get p...', 'path/pub...', 47, Array) #1 path/anchor/system/view.php(87): require('path/pub...') #2 path/anchor/system/view.php(72): System\View->render() #3 path/anchor/anchor/routes/pages.php(21): System\View->partial('header', 'partials/header') #4 [internal function]: {closure}() #5 path/anchor/system/route.php(152): call_user_func_array(Object(Closure), Array) #6 path/anchor/system/router.php(152): System\Route->run() #7 path/anchor/system/start.php(36): System\Router->dispatch() #8 path/anchor/index.php(33): require('path/pub...') #9 {main} ``` As it says, the problem is in anchor/views/partials/header.php line 47 ``` <?php echo Html::link($home->slug, __('global.visit_your_site'), array('class' => 'btn', 'target' => '_blank')); ?> ``` I was able to comment out that line and then access the admin and then set another page as a Home Page, but that was just a temporary fix. A solution would to be to restrict deleting a page, when it is set as Home Page.
1.0
Deleting a page that is assigned as Home Page makes the CMS unusable - I deleted a page (the default posts page) that was assigned as Home Page and after that I could not use the admin nor the page itself anymore because of: ``` Uncaught Exception Trying to get property of non-object Origin anchor/views/partials/header.php on line 47 Trace #0 path/anchor/anchor/views/partials/header.php(47): System\Error::native(8, 'Trying to get p...', 'path/pub...', 47, Array) #1 path/anchor/system/view.php(87): require('path/pub...') #2 path/anchor/system/view.php(72): System\View->render() #3 path/anchor/anchor/routes/pages.php(21): System\View->partial('header', 'partials/header') #4 [internal function]: {closure}() #5 path/anchor/system/route.php(152): call_user_func_array(Object(Closure), Array) #6 path/anchor/system/router.php(152): System\Route->run() #7 path/anchor/system/start.php(36): System\Router->dispatch() #8 path/anchor/index.php(33): require('path/pub...') #9 {main} ``` As it says, the problem is in anchor/views/partials/header.php line 47 ``` <?php echo Html::link($home->slug, __('global.visit_your_site'), array('class' => 'btn', 'target' => '_blank')); ?> ``` I was able to comment out that line and then access the admin and then set another page as a Home Page, but that was just a temporary fix. A solution would to be to restrict deleting a page, when it is set as Home Page.
non_process
deleting a page that is assigned as home page makes the cms unusable i deleted a page the default posts page that was assigned as home page and after that i could not use the admin nor the page itself anymore because of uncaught exception trying to get property of non object origin anchor views partials header php on line trace path anchor anchor views partials header php system error native trying to get p path pub array path anchor system view php require path pub path anchor system view php system view render path anchor anchor routes pages php system view partial header partials header closure path anchor system route php call user func array object closure array path anchor system router php system route run path anchor system start php system router dispatch path anchor index php require path pub main as it says the problem is in anchor views partials header php line slug global visit your site array class btn target blank i was able to comment out that line and then access the admin and then set another page as a home page but that was just a temporary fix a solution would to be to restrict deleting a page when it is set as home page
0
79,724
3,546,142,668
IssuesEvent
2016-01-20 00:42:36
18F/college-choice
https://api.github.com/repos/18F/college-choice
closed
As a product team member, I would like to see analytics for the consumer site so I can understand how visitors use it
Area - Consumer Tool Area - Data Tool Bang 2 - Med Bang For Buck 1 - Low Buck 2 - Med Priority 1 Scrub (Sabrina) Stage 6 - Accept
We'd like to have analytics before the beta so that we can see how beta users move through the site. I'm not sure if we can use [DAP](http://www.digitalgov.gov/services/dap/) for this, though, as I'm not sure what our top-level domain. I'm assigning this to @ultrasaurus pending decision on a hostname, as that will determine whether we can use DAP.
1.0
As a product team member, I would like to see analytics for the consumer site so I can understand how visitors use it - We'd like to have analytics before the beta so that we can see how beta users move through the site. I'm not sure if we can use [DAP](http://www.digitalgov.gov/services/dap/) for this, though, as I'm not sure what our top-level domain. I'm assigning this to @ultrasaurus pending decision on a hostname, as that will determine whether we can use DAP.
non_process
as a product team member i would like to see analytics for the consumer site so i can understand how visitors use it we d like to have analytics before the beta so that we can see how beta users move through the site i m not sure if we can use for this though as i m not sure what our top level domain i m assigning this to ultrasaurus pending decision on a hostname as that will determine whether we can use dap
0
20,534
27,190,744,808
IssuesEvent
2023-02-19 19:18:03
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
MIPS 32: Missing instructions
Type: Enhancement Feature: Processor/MIPS Status: Internal
**To Reproduce** The following byte sequences fail to disassemble in MIPS 32 (little endian): ``F8 FC 21 7C`` ``B8 0C 3F 7C`` ``01 00 60 41`` ``21 00 60 41`` **Expected behavior** Capstone (and IDA) returns: ``` <CsInsn 0x0 [f8fc217c]: wrdsp $at, 0x3f> <CsInsn 0x0 [b80c3f7c]: rddsp $at, 0x3f> ``` https://onlinedisassembler.com/odaweb/ (mips isa32r2) and IDA return ``` 01 00 60 41 dvpe 21 00 60 41 evpe ``` **Environment (please complete the following information):** Commit 2d51b95
1.0
MIPS 32: Missing instructions - **To Reproduce** The following byte sequences fail to disassemble in MIPS 32 (little endian): ``F8 FC 21 7C`` ``B8 0C 3F 7C`` ``01 00 60 41`` ``21 00 60 41`` **Expected behavior** Capstone (and IDA) returns: ``` <CsInsn 0x0 [f8fc217c]: wrdsp $at, 0x3f> <CsInsn 0x0 [b80c3f7c]: rddsp $at, 0x3f> ``` https://onlinedisassembler.com/odaweb/ (mips isa32r2) and IDA return ``` 01 00 60 41 dvpe 21 00 60 41 evpe ``` **Environment (please complete the following information):** Commit 2d51b95
process
mips missing instructions to reproduce the following byte sequences fail to disassemble in mips little endian fc expected behavior capstone and ida returns mips and ida return dvpe evpe environment please complete the following information commit
1
578,322
17,146,598,732
IssuesEvent
2021-07-13 15:12:00
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
appengine.flexible.tasks.snippets_test: test_retry_task failed
api: appengine flakybot: flaky flakybot: issue priority: p1 samples type: bug
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: b99df8d36109e4fe3e397bfd2cbacac06960340c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/325e6648-2604-4a2d-a611-53de8a0492aa), [Sponge](http://sponge2/325e6648-2604-4a2d-a611-53de8a0492aa) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 67, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." debug_error_string = "{"created":"@1626167050.183573535","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/snippets_test.py", line 127, in test_retry_task queue=QUEUE_NAME[i]) File "/workspace/appengine/flexible/tasks/snippets.py", line 233, in delete_queue response = client.delete_queue(name=queue_path) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/cloud/tasks_v2/services/cloud_tasks/client.py", line 815, in delete_queue request, retry=retry, timeout=timeout, metadata=metadata, File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py", line 290, in retry_wrapped_func on_error=on_error, File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py", line 188, in retry_target return target() File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 69, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.</pre></details>
1.0
appengine.flexible.tasks.snippets_test: test_retry_task failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: b99df8d36109e4fe3e397bfd2cbacac06960340c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/325e6648-2604-4a2d-a611-53de8a0492aa), [Sponge](http://sponge2/325e6648-2604-4a2d-a611-53de8a0492aa) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 67, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." debug_error_string = "{"created":"@1626167050.183573535","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/snippets_test.py", line 127, in test_retry_task queue=QUEUE_NAME[i]) File "/workspace/appengine/flexible/tasks/snippets.py", line 233, in delete_queue response = client.delete_queue(name=queue_path) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/cloud/tasks_v2/services/cloud_tasks/client.py", line 815, in delete_queue request, retry=retry, timeout=timeout, metadata=metadata, File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py", line 290, in retry_wrapped_func on_error=on_error, File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py", line 188, in retry_target return target() File "/workspace/appengine/flexible/tasks/.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py", line 69, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.</pre></details>
non_process
appengine flexible tasks snippets test test retry task failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace appengine flexible tasks nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace appengine flexible tasks nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace appengine flexible tasks nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unauthenticated details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see debug error string created description error received from peer file src core lib surface call cc file line grpc message request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see the above exception was the direct cause of the following exception traceback most recent call last file workspace appengine flexible tasks snippets test py line in test retry task queue queue name file workspace appengine flexible tasks snippets py line in delete queue response client delete queue name queue path file workspace appengine flexible tasks nox py lib site packages google cloud tasks services cloud tasks client py line in delete queue request retry retry timeout timeout metadata metadata file workspace appengine flexible tasks nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace appengine flexible tasks nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace appengine flexible tasks nox py lib site packages google api core retry py line in retry target return target file workspace appengine flexible tasks nox py lib site packages google api core grpc helpers py line in error remapped callable six raise from exceptions from grpc error exc exc file line in raise from google api core exceptions unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see
0
812,784
30,352,094,742
IssuesEvent
2023-07-11 19:49:48
kubernetes/release
https://api.github.com/repos/kubernetes/release
closed
Dependency update - Golang 1.20.5/1.19.10
kind/feature priority/important-soon sig/release area/release-eng area/dependency
<!-- Please only use this template if you are a Release Manager updating Golang dependencies. --> ### Tracking info <!-- Search query: https://github.com/kubernetes/release/issues?q=is%3Aissue+Dependency+update+-+Golang --> <!-- Example: https://github.com/kubernetes/release/issues/2694 --> Link to any previous tracking issue: https://github.com/kubernetes/release/issues/3025 <!-- golang-announce mailing list: https://groups.google.com/g/golang-announce --> Golang mailing list announcement: https://groups.google.com/g/golang-announce/c/1AItFMBjrfw SIG Release Slack thread: https://kubernetes.slack.com/archives/CJH2GBF7Y/p1685659171113989 ### Work items <!-- Example: https://github.com/kubernetes/release/pull/2696 --> - [x] `kube-cross` image update: https://github.com/kubernetes/release/pull/3105 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4312 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5375 <!-- Example: https://github.com/kubernetes/release/pull/2696 --> - [x] `go-runner` image update: https://github.com/kubernetes/release/pull/3105 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4313 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5376 <!-- Example: https://github.com/kubernetes/release/pull/2696 --> - [x] `releng-ci` image update: https://github.com/kubernetes/release/pull/3105 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4314 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5377 #### After kube-cross image promotion <!-- Example: https://github.com/kubernetes/kubernetes/pull/112900 --> - [x] kubernetes/kubernetes update (`master`): https://github.com/kubernetes/kubernetes/pull/118507 Ensure the following have been updated within the PR: - [x] kube-cross image - [x] go-runner image - [x] publishing bot rules - [x] test image - [x] `.go-version` file #### After go-runner image promotion <!-- Example: https://github.com/kubernetes/release/pull/2920 --> - [x] `debian-iptables` image update: https://github.com/kubernetes/release/pull/3107 <!-- Example: https://github.com/kubernetes/k8s.io/pull/3597 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5383 <!-- Example: https://github.com/kubernetes/release/pull/2920 --> - [x] `distroless-iptables` image update: https://github.com/kubernetes/release/pull/3107 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4263 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5383 - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5384 #### After distroless-iptables image promotion <!-- Example: https://github.com/kubernetes/kubernetes/pull/112892 --> - [x] kubernetes/kubernetes update (`master`): https://github.com/kubernetes/kubernetes/pull/118541 Ensure the following have been updated within the PR: - [x] distroless-iptables image - [x] test image #### After kubernetes/kubernetes (master) has been updated <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-cloud-builder` image update: https://github.com/kubernetes/release/pull/3108 <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-ci-builder` image variants update: https://github.com/kubernetes/release/pull/3108 ### Cherry picks <!-- Depending on the Golang release type, this section may not be required. General rule of thumb: Only cherry pick Golang patch releases to branches that have the same Golang minor release version. Concrete example: At the time of this template's creation, go1.15.5 was just merged on our primary development branch and the following Golang versions were active on in-support kubernetes/kubernetes release branches: - `master`: go1.15.5 - `release-1.19`: go1.15.2 - `release-1.18`: go1.13.15 - `release-1.17`: go1.13.15 In this case, we would only cherry pick the go1.15.5 to the `release-1.19` branch, since it is the only other branch with a go1.15 minor version on it. --> - [x] Kubernetes 1.27: https://github.com/kubernetes/kubernetes/pull/118553 - [x] Kubernetes 1.26: https://github.com/kubernetes/kubernetes/pull/118555 - [x] Kubernetes 1.25: https://github.com/kubernetes/kubernetes/pull/118556 - [x] Kubernetes 1.24: https://github.com/kubernetes/kubernetes/pull/118557 <!-- If the Golang version of the active development branch (`master`) is newer than any of the Golang versions on _active_ release branches, then the current Golang versions for all release branches need to be updated within publishing bot rules. Concrete example: - `master` was just updated from go1.16.6 to go1.16.7 - cherry picks were issued to the 1.22 and 1.21 branches - `release-1.20` was also updated from go1.15.14 to go1.15.15 - these changes were cherry picked to the 1.19 branch In this case, because we updated the default go version on `master` to go1.16.7, there's no action required for staging repositories using go1.16. However, for staging repository branches using go1.15, the `master` branch's publishing bot rules need to be updated to learn about the Golang update that happened for the 1.20 and 1.19 Kubernetes release branches. PR: https://github.com/kubernetes/kubernetes/pull/104226 --> - [x] publishing bot rule updates for active Golang versions: https://github.com/kubernetes/kubernetes/pull/118629 #### After kubernetes/kubernetes (release branches) has been updated <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-cloud-builder` image update: https://github.com/kubernetes/release/pull/3116 <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-ci-builder` image variants update: https://github.com/kubernetes/release/pull/3116 <!-- Example: https://github.com/kubernetes/test-infra/pull/27712 --> - [x] `kubekins`/`krte` image variants update: https://github.com/kubernetes/test-infra/pull/29781 ### Follow-up items <!-- Use this section to list out process improvements or items that need to be addressed before the next Golang update. --> - [ ] Ensure the Golang issue template is updated with any new requirements - [ ] <Any other follow up items> /assign cc: @kubernetes/release-engineering
1.0
Dependency update - Golang 1.20.5/1.19.10 - <!-- Please only use this template if you are a Release Manager updating Golang dependencies. --> ### Tracking info <!-- Search query: https://github.com/kubernetes/release/issues?q=is%3Aissue+Dependency+update+-+Golang --> <!-- Example: https://github.com/kubernetes/release/issues/2694 --> Link to any previous tracking issue: https://github.com/kubernetes/release/issues/3025 <!-- golang-announce mailing list: https://groups.google.com/g/golang-announce --> Golang mailing list announcement: https://groups.google.com/g/golang-announce/c/1AItFMBjrfw SIG Release Slack thread: https://kubernetes.slack.com/archives/CJH2GBF7Y/p1685659171113989 ### Work items <!-- Example: https://github.com/kubernetes/release/pull/2696 --> - [x] `kube-cross` image update: https://github.com/kubernetes/release/pull/3105 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4312 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5375 <!-- Example: https://github.com/kubernetes/release/pull/2696 --> - [x] `go-runner` image update: https://github.com/kubernetes/release/pull/3105 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4313 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5376 <!-- Example: https://github.com/kubernetes/release/pull/2696 --> - [x] `releng-ci` image update: https://github.com/kubernetes/release/pull/3105 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4314 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5377 #### After kube-cross image promotion <!-- Example: https://github.com/kubernetes/kubernetes/pull/112900 --> - [x] kubernetes/kubernetes update (`master`): https://github.com/kubernetes/kubernetes/pull/118507 Ensure the following have been updated within the PR: - [x] kube-cross image - [x] go-runner image - [x] publishing bot rules - [x] test image - [x] `.go-version` file #### After go-runner image promotion <!-- Example: https://github.com/kubernetes/release/pull/2920 --> - [x] `debian-iptables` image update: https://github.com/kubernetes/release/pull/3107 <!-- Example: https://github.com/kubernetes/k8s.io/pull/3597 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5383 <!-- Example: https://github.com/kubernetes/release/pull/2920 --> - [x] `distroless-iptables` image update: https://github.com/kubernetes/release/pull/3107 <!-- Example: https://github.com/kubernetes/k8s.io/pull/4263 --> - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5383 - [x] image promotion: https://github.com/kubernetes/k8s.io/pull/5384 #### After distroless-iptables image promotion <!-- Example: https://github.com/kubernetes/kubernetes/pull/112892 --> - [x] kubernetes/kubernetes update (`master`): https://github.com/kubernetes/kubernetes/pull/118541 Ensure the following have been updated within the PR: - [x] distroless-iptables image - [x] test image #### After kubernetes/kubernetes (master) has been updated <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-cloud-builder` image update: https://github.com/kubernetes/release/pull/3108 <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-ci-builder` image variants update: https://github.com/kubernetes/release/pull/3108 ### Cherry picks <!-- Depending on the Golang release type, this section may not be required. General rule of thumb: Only cherry pick Golang patch releases to branches that have the same Golang minor release version. Concrete example: At the time of this template's creation, go1.15.5 was just merged on our primary development branch and the following Golang versions were active on in-support kubernetes/kubernetes release branches: - `master`: go1.15.5 - `release-1.19`: go1.15.2 - `release-1.18`: go1.13.15 - `release-1.17`: go1.13.15 In this case, we would only cherry pick the go1.15.5 to the `release-1.19` branch, since it is the only other branch with a go1.15 minor version on it. --> - [x] Kubernetes 1.27: https://github.com/kubernetes/kubernetes/pull/118553 - [x] Kubernetes 1.26: https://github.com/kubernetes/kubernetes/pull/118555 - [x] Kubernetes 1.25: https://github.com/kubernetes/kubernetes/pull/118556 - [x] Kubernetes 1.24: https://github.com/kubernetes/kubernetes/pull/118557 <!-- If the Golang version of the active development branch (`master`) is newer than any of the Golang versions on _active_ release branches, then the current Golang versions for all release branches need to be updated within publishing bot rules. Concrete example: - `master` was just updated from go1.16.6 to go1.16.7 - cherry picks were issued to the 1.22 and 1.21 branches - `release-1.20` was also updated from go1.15.14 to go1.15.15 - these changes were cherry picked to the 1.19 branch In this case, because we updated the default go version on `master` to go1.16.7, there's no action required for staging repositories using go1.16. However, for staging repository branches using go1.15, the `master` branch's publishing bot rules need to be updated to learn about the Golang update that happened for the 1.20 and 1.19 Kubernetes release branches. PR: https://github.com/kubernetes/kubernetes/pull/104226 --> - [x] publishing bot rule updates for active Golang versions: https://github.com/kubernetes/kubernetes/pull/118629 #### After kubernetes/kubernetes (release branches) has been updated <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-cloud-builder` image update: https://github.com/kubernetes/release/pull/3116 <!-- Example: https://github.com/kubernetes/release/pull/2699 --> - [x] `k8s-ci-builder` image variants update: https://github.com/kubernetes/release/pull/3116 <!-- Example: https://github.com/kubernetes/test-infra/pull/27712 --> - [x] `kubekins`/`krte` image variants update: https://github.com/kubernetes/test-infra/pull/29781 ### Follow-up items <!-- Use this section to list out process improvements or items that need to be addressed before the next Golang update. --> - [ ] Ensure the Golang issue template is updated with any new requirements - [ ] <Any other follow up items> /assign cc: @kubernetes/release-engineering
non_process
dependency update golang please only use this template if you are a release manager updating golang dependencies tracking info link to any previous tracking issue golang mailing list announcement sig release slack thread work items kube cross image update image promotion go runner image update image promotion releng ci image update image promotion after kube cross image promotion kubernetes kubernetes update master ensure the following have been updated within the pr kube cross image go runner image publishing bot rules test image go version file after go runner image promotion debian iptables image update image promotion distroless iptables image update image promotion image promotion after distroless iptables image promotion kubernetes kubernetes update master ensure the following have been updated within the pr distroless iptables image test image after kubernetes kubernetes master has been updated cloud builder image update ci builder image variants update cherry picks depending on the golang release type this section may not be required general rule of thumb only cherry pick golang patch releases to branches that have the same golang minor release version concrete example at the time of this template s creation was just merged on our primary development branch and the following golang versions were active on in support kubernetes kubernetes release branches master release release release in this case we would only cherry pick the to the release branch since it is the only other branch with a minor version on it kubernetes kubernetes kubernetes kubernetes if the golang version of the active development branch master is newer than any of the golang versions on active release branches then the current golang versions for all release branches need to be updated within publishing bot rules concrete example master was just updated from to cherry picks were issued to the and branches release was also updated from to these changes were cherry picked to the branch in this case because we updated the default go version on master to there s no action required for staging repositories using however for staging repository branches using the master branch s publishing bot rules need to be updated to learn about the golang update that happened for the and kubernetes release branches pr publishing bot rule updates for active golang versions after kubernetes kubernetes release branches has been updated cloud builder image update ci builder image variants update kubekins krte image variants update follow up items use this section to list out process improvements or items that need to be addressed before the next golang update ensure the golang issue template is updated with any new requirements assign cc kubernetes release engineering
0
6,523
9,608,728,346
IssuesEvent
2019-05-12 09:15:56
shirou/gopsutil
https://api.github.com/repos/shirou/gopsutil
closed
Race condition in `process.Exe()` on macOS
os:darwin package:process
**Describe the bug** A race condition triggers when calling the `Exe()` method on macOS: https://github.com/shirou/gopsutil/blob/master/process/process_darwin.go#L102 ``` ================== WARNING: DATA RACE Write at 0x00c0000b25f0 by goroutine 8: bytes.(*Buffer).ReadFrom() /usr/local/go/src/bytes/buffer.go:203 +0x4b io.copyBuffer() /usr/local/go/src/io/io.go:388 +0x3f5 os/exec.(*Cmd).writerDescriptor.func1() /usr/local/go/src/io/io.go:364 +0x7a os/exec.(*Cmd).Start.func1() /usr/local/go/src/os/exec/exec.go:409 +0x34 Previous write at 0x00c0000b25f0 by goroutine 6: bytes.(*Buffer).ReadFrom() /usr/local/go/src/bytes/buffer.go:203 +0x4b io.copyBuffer() /usr/local/go/src/io/io.go:388 +0x3f5 os/exec.(*Cmd).writerDescriptor.func1() /usr/local/go/src/io/io.go:364 +0x7a os/exec.(*Cmd).Start.func1() /usr/local/go/src/os/exec/exec.go:409 +0x34 Goroutine 8 (running) created at: os/exec.(*Cmd).Start() /usr/local/go/src/os/exec/exec.go:408 +0x9c3 github.com/shirou/gopsutil/internal/common.Pipeline() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/internal/common/common.go:362 +0x477 github.com/shirou/gopsutil/process.(*Process).ExeWithContext() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:126 +0x40d main.main() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:103 +0xcd Goroutine 6 (running) created at: os/exec.(*Cmd).Start() /usr/local/go/src/os/exec/exec.go:408 +0x9c3 github.com/shirou/gopsutil/internal/common.Pipeline() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/internal/common/common.go:362 +0x477 github.com/shirou/gopsutil/process.(*Process).ExeWithContext() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:126 +0x40d main.main() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:103 +0xcd ================== ``` **To Reproduce** Replace `<insert pid>` with a valid PID from your system and build using `go build -race`. Running it will immediately trigger a race condition. ``` package main import ( "github.com/shirou/gopsutil/process" ) func main() { procData, _ := process.NewProcess(<insert pid>) procData.Exe() } ``` **Expected behavior** A race condition should not occur. **Environment (please complete the following information):** - [ ] Windows: [paste the result of `ver`] - [ ] Linux: [paste contents of `/etc/os-release` and the result of `uname -a`] - [ ] Mac OS: [paste the result of `sw_vers` and `uname -a` - [ ] FreeBSD: [paste the result of `freebsd-version -k -r -u` and `uname -a`] - [ ] OpenBSD: [paste the result of `uname -a`] ``` ProductName: Mac OS X ProductVersion: 10.14.3 BuildVersion: 18D109 Darwin Brians-MacBook-Pro.local 18.2.0 Darwin Kernel Version 18.2.0: Thu Dec 20 20:46:53 PST 2018; root:xnu-4903.241.1~1/RELEASE_X86_64 x86_64 ``` **Additional context** [Cross-compiling? Paste the command you are using to cross-compile and the result of the corresponding `go env`]
1.0
Race condition in `process.Exe()` on macOS - **Describe the bug** A race condition triggers when calling the `Exe()` method on macOS: https://github.com/shirou/gopsutil/blob/master/process/process_darwin.go#L102 ``` ================== WARNING: DATA RACE Write at 0x00c0000b25f0 by goroutine 8: bytes.(*Buffer).ReadFrom() /usr/local/go/src/bytes/buffer.go:203 +0x4b io.copyBuffer() /usr/local/go/src/io/io.go:388 +0x3f5 os/exec.(*Cmd).writerDescriptor.func1() /usr/local/go/src/io/io.go:364 +0x7a os/exec.(*Cmd).Start.func1() /usr/local/go/src/os/exec/exec.go:409 +0x34 Previous write at 0x00c0000b25f0 by goroutine 6: bytes.(*Buffer).ReadFrom() /usr/local/go/src/bytes/buffer.go:203 +0x4b io.copyBuffer() /usr/local/go/src/io/io.go:388 +0x3f5 os/exec.(*Cmd).writerDescriptor.func1() /usr/local/go/src/io/io.go:364 +0x7a os/exec.(*Cmd).Start.func1() /usr/local/go/src/os/exec/exec.go:409 +0x34 Goroutine 8 (running) created at: os/exec.(*Cmd).Start() /usr/local/go/src/os/exec/exec.go:408 +0x9c3 github.com/shirou/gopsutil/internal/common.Pipeline() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/internal/common/common.go:362 +0x477 github.com/shirou/gopsutil/process.(*Process).ExeWithContext() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:126 +0x40d main.main() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:103 +0xcd Goroutine 6 (running) created at: os/exec.(*Cmd).Start() /usr/local/go/src/os/exec/exec.go:408 +0x9c3 github.com/shirou/gopsutil/internal/common.Pipeline() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/internal/common/common.go:362 +0x477 github.com/shirou/gopsutil/process.(*Process).ExeWithContext() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:126 +0x40d main.main() /Users/bluger/dev/go/pkg/mod/github.com/shirou/gopsutil@v2.18.12+incompatible/process/process_darwin.go:103 +0xcd ================== ``` **To Reproduce** Replace `<insert pid>` with a valid PID from your system and build using `go build -race`. Running it will immediately trigger a race condition. ``` package main import ( "github.com/shirou/gopsutil/process" ) func main() { procData, _ := process.NewProcess(<insert pid>) procData.Exe() } ``` **Expected behavior** A race condition should not occur. **Environment (please complete the following information):** - [ ] Windows: [paste the result of `ver`] - [ ] Linux: [paste contents of `/etc/os-release` and the result of `uname -a`] - [ ] Mac OS: [paste the result of `sw_vers` and `uname -a` - [ ] FreeBSD: [paste the result of `freebsd-version -k -r -u` and `uname -a`] - [ ] OpenBSD: [paste the result of `uname -a`] ``` ProductName: Mac OS X ProductVersion: 10.14.3 BuildVersion: 18D109 Darwin Brians-MacBook-Pro.local 18.2.0 Darwin Kernel Version 18.2.0: Thu Dec 20 20:46:53 PST 2018; root:xnu-4903.241.1~1/RELEASE_X86_64 x86_64 ``` **Additional context** [Cross-compiling? Paste the command you are using to cross-compile and the result of the corresponding `go env`]
process
race condition in process exe on macos describe the bug a race condition triggers when calling the exe method on macos warning data race write at by goroutine bytes buffer readfrom usr local go src bytes buffer go io copybuffer usr local go src io io go os exec cmd writerdescriptor usr local go src io io go os exec cmd start usr local go src os exec exec go previous write at by goroutine bytes buffer readfrom usr local go src bytes buffer go io copybuffer usr local go src io io go os exec cmd writerdescriptor usr local go src io io go os exec cmd start usr local go src os exec exec go goroutine running created at os exec cmd start usr local go src os exec exec go github com shirou gopsutil internal common pipeline users bluger dev go pkg mod github com shirou gopsutil incompatible internal common common go github com shirou gopsutil process process exewithcontext users bluger dev go pkg mod github com shirou gopsutil incompatible process process darwin go main main users bluger dev go pkg mod github com shirou gopsutil incompatible process process darwin go goroutine running created at os exec cmd start usr local go src os exec exec go github com shirou gopsutil internal common pipeline users bluger dev go pkg mod github com shirou gopsutil incompatible internal common common go github com shirou gopsutil process process exewithcontext users bluger dev go pkg mod github com shirou gopsutil incompatible process process darwin go main main users bluger dev go pkg mod github com shirou gopsutil incompatible process process darwin go to reproduce replace with a valid pid from your system and build using go build race running it will immediately trigger a race condition package main import github com shirou gopsutil process func main procdata process newprocess procdata exe expected behavior a race condition should not occur environment please complete the following information windows linux mac os paste the result of sw vers and uname a freebsd openbsd productname mac os x productversion buildversion darwin brians macbook pro local darwin kernel version thu dec pst root xnu release additional context
1
553
3,014,162,417
IssuesEvent
2015-07-29 13:35:46
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
На дашборде подвязать использоване сервиса экспорта тасок по БП ("downloadTasksData"), в разделе "Звіт"
hi priority In process of testing question test
На дашборде, если у пользователь состоит в группе "mngr": (это можно отрисовать сразу в отдельном разделе, как в таске: https://github.com/e-government-ua/i/issues/560) 1) отрисовать поле input для даты (умолчательно ставить прошлый день) 2) отрисовать кнопку "Экспортировать", при нажатии на которую вызывать сервис "downloadTasksData" (параметры слать те-же что в таске 517, только дату из п.1) 3) параметры для сервиса передавать с клиента а не сервера (позже будет комбобокс с выбором вариантов) Сервис "downloadTasksData" реализовывался по таске: https://github.com/e-government-ua/i/issues/517 П.С.: даже если сервис окончательно не оттестирован - привязаться "даже без особых тестов" (главное чтоб вызывался сервис, а он уже точно будет предлагать сохранять файл)
1.0
На дашборде подвязать использоване сервиса экспорта тасок по БП ("downloadTasksData"), в разделе "Звіт" - На дашборде, если у пользователь состоит в группе "mngr": (это можно отрисовать сразу в отдельном разделе, как в таске: https://github.com/e-government-ua/i/issues/560) 1) отрисовать поле input для даты (умолчательно ставить прошлый день) 2) отрисовать кнопку "Экспортировать", при нажатии на которую вызывать сервис "downloadTasksData" (параметры слать те-же что в таске 517, только дату из п.1) 3) параметры для сервиса передавать с клиента а не сервера (позже будет комбобокс с выбором вариантов) Сервис "downloadTasksData" реализовывался по таске: https://github.com/e-government-ua/i/issues/517 П.С.: даже если сервис окончательно не оттестирован - привязаться "даже без особых тестов" (главное чтоб вызывался сервис, а он уже точно будет предлагать сохранять файл)
process
на дашборде подвязать использоване сервиса экспорта тасок по бп downloadtasksdata в разделе звіт на дашборде если у пользователь состоит в группе mngr это можно отрисовать сразу в отдельном разделе как в таске отрисовать поле input для даты умолчательно ставить прошлый день отрисовать кнопку экспортировать при нажатии на которую вызывать сервис downloadtasksdata параметры слать те же что в таске только дату из п параметры для сервиса передавать с клиента а не сервера позже будет комбобокс с выбором вариантов сервис downloadtasksdata реализовывался по таске п с даже если сервис окончательно не оттестирован привязаться даже без особых тестов главное чтоб вызывался сервис а он уже точно будет предлагать сохранять файл
1
4,725
2,742,121,008
IssuesEvent
2015-04-21 15:03:16
softlayer/sl-ember-components
https://api.github.com/repos/softlayer/sl-ember-components
closed
Add tests for sl-calendar component
0 - Backlog sl-calendar tests
<!--- @huboard:{"order":1.7053025658242404e-13,"milestone_order":206,"custom_state":""} -->
1.0
Add tests for sl-calendar component - <!--- @huboard:{"order":1.7053025658242404e-13,"milestone_order":206,"custom_state":""} -->
non_process
add tests for sl calendar component huboard order milestone order custom state
0
13,914
16,674,483,595
IssuesEvent
2021-06-07 14:40:58
Ayush7614/Daily-Coding-DS-ALGO-Practice
https://api.github.com/repos/Ayush7614/Daily-Coding-DS-ALGO-Practice
closed
Binary Search in c
Medium discussion enhancement good first issue question under-process
Hey @Ayush7614, I wanted to add binary search in c Please assign me to do it... Thanks
1.0
Binary Search in c - Hey @Ayush7614, I wanted to add binary search in c Please assign me to do it... Thanks
process
binary search in c hey i wanted to add binary search in c please assign me to do it thanks
1
1,286
3,534,633,256
IssuesEvent
2016-01-16 00:16:17
BCDevExchange/Our-Project-Docs
https://api.github.com/repos/BCDevExchange/Our-Project-Docs
opened
API List Page - API console request label
API Services
Where a list of requests for an API does not have a label (default), replace with a label 'resources'. Updates needed to: > Kamloops - DONE > Open511 - INCOMPLETE > BCDC - INCOMPLETE
1.0
API List Page - API console request label - Where a list of requests for an API does not have a label (default), replace with a label 'resources'. Updates needed to: > Kamloops - DONE > Open511 - INCOMPLETE > BCDC - INCOMPLETE
non_process
api list page api console request label where a list of requests for an api does not have a label default replace with a label resources updates needed to kamloops done incomplete bcdc incomplete
0
83,849
15,720,737,353
IssuesEvent
2021-03-29 01:02:43
vlaship/build-docker-image
https://api.github.com/repos/vlaship/build-docker-image
opened
CVE-2021-24122 (Medium) detected in tomcat-embed-core-9.0.26.jar
security vulnerability
## CVE-2021-24122 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.26.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: build-docker-image/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.9.RELEASE.jar - tomcat-embed-websocket-9.0.26.jar - :x: **tomcat-embed-core-9.0.26.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances. <p>Publish Date: 2021-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122>CVE-2021-24122</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122</a></p> <p>Release Date: 2021-01-14</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-24122 (Medium) detected in tomcat-embed-core-9.0.26.jar - ## CVE-2021-24122 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.26.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: build-docker-image/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.9.RELEASE.jar - tomcat-embed-websocket-9.0.26.jar - :x: **tomcat-embed-core-9.0.26.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When serving resources from a network location using the NTFS file system, Apache Tomcat versions 10.0.0-M1 to 10.0.0-M9, 9.0.0.M1 to 9.0.39, 8.5.0 to 8.5.59 and 7.0.0 to 7.0.106 were susceptible to JSP source code disclosure in some configurations. The root cause was the unexpected behaviour of the JRE API File.getCanonicalPath() which in turn was caused by the inconsistent behaviour of the Windows API (FindFirstFileW) in some circumstances. <p>Publish Date: 2021-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24122>CVE-2021-24122</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-24122</a></p> <p>Release Date: 2021-01-14</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.107,8.5.60,9.0.40,10.0.0-M10;org.apache.tomcat:tomcat-catalina:7.0.107,8.5.60,9.0.40,10.0.0-M10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in tomcat embed core jar cve medium severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file build docker image build gradle path to vulnerable library root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar tomcat embed websocket jar x tomcat embed core jar vulnerable library vulnerability details when serving resources from a network location using the ntfs file system apache tomcat versions to to to and to were susceptible to jsp source code disclosure in some configurations the root cause was the unexpected behaviour of the jre api file getcanonicalpath which in turn was caused by the inconsistent behaviour of the windows api findfirstfilew in some circumstances publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
0
136,362
19,768,282,512
IssuesEvent
2022-01-17 06:59:40
5balls/coleitra
https://api.github.com/repos/5balls/coleitra
closed
Make grammarprovider configurable
enhancement design functionality
Implement all stuff needed to just use json files to configure / extend different language support in grammarprovider - [ ] Make reasonable data structure - [ ] Load json files found into such list of data structures - [ ] Implement calling function mechanism for this data structure
1.0
Make grammarprovider configurable - Implement all stuff needed to just use json files to configure / extend different language support in grammarprovider - [ ] Make reasonable data structure - [ ] Load json files found into such list of data structures - [ ] Implement calling function mechanism for this data structure
non_process
make grammarprovider configurable implement all stuff needed to just use json files to configure extend different language support in grammarprovider make reasonable data structure load json files found into such list of data structures implement calling function mechanism for this data structure
0
11,614
14,480,396,597
IssuesEvent
2020-12-10 11:06:58
cggos/cggos.github.io
https://api.github.com/repos/cggos/cggos.github.io
opened
图像分析之图像特征匹配 - Gavin Gao's Blog
Gitalk image-process-feature-matching
https://cggos.github.io/image-process-feature-matching.html [TOC]相似度 SSD (Sum of Squared Distance)\[{D(I_A,I_B)}_{SSD} = \sum_{x,y}[{I_A}(x,y)-{I_B}(x,y)]^2\] SAD (Sum of Absolute Difference)\[{D(I_A,I_B)}_{SAD} = \...
1.0
图像分析之图像特征匹配 - Gavin Gao's Blog - https://cggos.github.io/image-process-feature-matching.html [TOC]相似度 SSD (Sum of Squared Distance)\[{D(I_A,I_B)}_{SSD} = \sum_{x,y}[{I_A}(x,y)-{I_B}(x,y)]^2\] SAD (Sum of Absolute Difference)\[{D(I_A,I_B)}_{SAD} = \...
process
图像分析之图像特征匹配 gavin gao s blog 相似度 ssd sum of squared distance sad sum of absolute difference d i a i b sad
1
17,800
23,728,190,900
IssuesEvent
2022-08-30 21:53:45
GSA/EDX
https://api.github.com/repos/GSA/EDX
closed
OMB orphan sites
process policy strategy executive board collaboration
## Summary GSA is often tasked by OMB to stand up a new website for certain initiatives or policies. Unfortunately, OMB is not great at maintaining these sites over time, so GSA is left with a lot of sites that were commissioned by OMB, but are no longer supported. We need a way to communicate with OMB about cleaning up/closing/consolidating these sites, and take them down, or resource them. [Analysis of MY sites](https://docs.google.com/document/d/1BL0m4FIsG1EgAtxnfNtOAbnq_HaZR9bm_7DTrc0SLmE/edit#heading=h.1nvioducx0jp) ## Checklist - [x] Identify which websites fit this category (e.g., some *.data.gov and *.cio.gov sites) - [x] Meet with OGP/MY to discuss - [x] Determine which sites are actually under GSA control versus those which occupy domain name space (OGP is simply the "putter upper" and has no control over content) - [x] Update inventory to recategorize OMB sites as "external" brand category - [x] Determine with OGP whether/how to raise this issue w/OMB - [x] Communicate with OMB about this (email from Earl) - [x] Advise OGP to remove GSA branding from "external" sites ## List of Sites ### GSA Control |URL| Github Repo| |---|------------| |[federation.data.gov](https://federation.data.gov)|[github.com/GSA/us-data-federation](https://github.com/GSA/us-data-federation)| ### OMB Manages Repository |URL| Github Repo| |---|------------| |[management.cio.gov/](https://management.cio.gov/)|[github.com/WhiteHouse/CIOmanagement](https://github.com/WhiteHouse/CIOmanagement)| |[software.cio.gov](https://software.cio.gov/)|[github.com/WhiteHouse/software-policy](https://github.com/WhiteHouse/software-policy)|
1.0
OMB orphan sites - ## Summary GSA is often tasked by OMB to stand up a new website for certain initiatives or policies. Unfortunately, OMB is not great at maintaining these sites over time, so GSA is left with a lot of sites that were commissioned by OMB, but are no longer supported. We need a way to communicate with OMB about cleaning up/closing/consolidating these sites, and take them down, or resource them. [Analysis of MY sites](https://docs.google.com/document/d/1BL0m4FIsG1EgAtxnfNtOAbnq_HaZR9bm_7DTrc0SLmE/edit#heading=h.1nvioducx0jp) ## Checklist - [x] Identify which websites fit this category (e.g., some *.data.gov and *.cio.gov sites) - [x] Meet with OGP/MY to discuss - [x] Determine which sites are actually under GSA control versus those which occupy domain name space (OGP is simply the "putter upper" and has no control over content) - [x] Update inventory to recategorize OMB sites as "external" brand category - [x] Determine with OGP whether/how to raise this issue w/OMB - [x] Communicate with OMB about this (email from Earl) - [x] Advise OGP to remove GSA branding from "external" sites ## List of Sites ### GSA Control |URL| Github Repo| |---|------------| |[federation.data.gov](https://federation.data.gov)|[github.com/GSA/us-data-federation](https://github.com/GSA/us-data-federation)| ### OMB Manages Repository |URL| Github Repo| |---|------------| |[management.cio.gov/](https://management.cio.gov/)|[github.com/WhiteHouse/CIOmanagement](https://github.com/WhiteHouse/CIOmanagement)| |[software.cio.gov](https://software.cio.gov/)|[github.com/WhiteHouse/software-policy](https://github.com/WhiteHouse/software-policy)|
process
omb orphan sites summary gsa is often tasked by omb to stand up a new website for certain initiatives or policies unfortunately omb is not great at maintaining these sites over time so gsa is left with a lot of sites that were commissioned by omb but are no longer supported we need a way to communicate with omb about cleaning up closing consolidating these sites and take them down or resource them checklist identify which websites fit this category e g some data gov and cio gov sites meet with ogp my to discuss determine which sites are actually under gsa control versus those which occupy domain name space ogp is simply the putter upper and has no control over content update inventory to recategorize omb sites as external brand category determine with ogp whether how to raise this issue w omb communicate with omb about this email from earl advise ogp to remove gsa branding from external sites list of sites gsa control url github repo omb manages repository url github repo
1
9,516
12,498,527,511
IssuesEvent
2020-06-01 18:27:04
w3c/webauthn
https://api.github.com/repos/w3c/webauthn
closed
MDN panels do not appear in editors' draft
type:process
pr #1411 landed, but the "MDN Panels" are not appearing in the editors' draft https://w3c.github.io/webauthn/ --- however, they do appear when I build the spec locally :(
1.0
MDN panels do not appear in editors' draft - pr #1411 landed, but the "MDN Panels" are not appearing in the editors' draft https://w3c.github.io/webauthn/ --- however, they do appear when I build the spec locally :(
process
mdn panels do not appear in editors draft pr landed but the mdn panels are not appearing in the editors draft however they do appear when i build the spec locally
1
232,861
25,706,381,970
IssuesEvent
2022-12-07 01:07:50
nidhi7598/linux-4.1.15_CVE-2019-10220
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2019-10220
opened
CVE-2017-14340 (Medium) detected in linuxlinux-4.6
security vulnerability
## CVE-2017-14340 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/xfs/xfs_linux.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The XFS_IS_REALTIME_INODE macro in fs/xfs/xfs_linux.h in the Linux kernel before 4.13.2 does not verify that a filesystem has a realtime device, which allows local users to cause a denial of service (NULL pointer dereference and OOPS) via vectors related to setting an RHINHERIT flag on a directory. <p>Publish Date: 2017-09-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-14340>CVE-2017-14340</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-14340">https://nvd.nist.gov/vuln/detail/CVE-2017-14340</a></p> <p>Release Date: 2017-09-15</p> <p>Fix Resolution: 4.13.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-14340 (Medium) detected in linuxlinux-4.6 - ## CVE-2017-14340 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/xfs/xfs_linux.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The XFS_IS_REALTIME_INODE macro in fs/xfs/xfs_linux.h in the Linux kernel before 4.13.2 does not verify that a filesystem has a realtime device, which allows local users to cause a denial of service (NULL pointer dereference and OOPS) via vectors related to setting an RHINHERIT flag on a directory. <p>Publish Date: 2017-09-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-14340>CVE-2017-14340</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-14340">https://nvd.nist.gov/vuln/detail/CVE-2017-14340</a></p> <p>Release Date: 2017-09-15</p> <p>Fix Resolution: 4.13.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files fs xfs xfs linux h vulnerability details the xfs is realtime inode macro in fs xfs xfs linux h in the linux kernel before does not verify that a filesystem has a realtime device which allows local users to cause a denial of service null pointer dereference and oops via vectors related to setting an rhinherit flag on a directory publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
9,602
12,544,431,219
IssuesEvent
2020-06-05 17:11:22
near/devx
https://api.github.com/repos/near/devx
closed
Process for responding DevX questions
P1 process
Do we create issues every time we get a question like this https://near-protocol.slack.com/archives/C01346BPS5T/p1589656379104800 ? E.g. there are 3 issues here: - AS sdk doesn’t have a high level API for funds transfer - We don’t have an example of AS contract transferring NEAR :face_palm: - We don’t have documentation for transferring NEAR (in both AS and RS) But more specifically, it’s the process of capturing these -> e.g. ideally the discussion above should be concluded with 3 issues linked. Where 3rd issue should contain response to developer (e.g. like this - https://github.com/nearprotocol/docs/issues/341)
1.0
Process for responding DevX questions - Do we create issues every time we get a question like this https://near-protocol.slack.com/archives/C01346BPS5T/p1589656379104800 ? E.g. there are 3 issues here: - AS sdk doesn’t have a high level API for funds transfer - We don’t have an example of AS contract transferring NEAR :face_palm: - We don’t have documentation for transferring NEAR (in both AS and RS) But more specifically, it’s the process of capturing these -> e.g. ideally the discussion above should be concluded with 3 issues linked. Where 3rd issue should contain response to developer (e.g. like this - https://github.com/nearprotocol/docs/issues/341)
process
process for responding devx questions do we create issues every time we get a question like this e g there are issues here as sdk doesn’t have a high level api for funds transfer we don’t have an example of as contract transferring near face palm we don’t have documentation for transferring near in both as and rs but more specifically it’s the process of capturing these e g ideally the discussion above should be concluded with issues linked where issue should contain response to developer e g like this
1
413,779
27,969,134,930
IssuesEvent
2023-03-24 23:03:05
mother-of-all-self-hosting/ansible-role-gotosocial
https://api.github.com/repos/mother-of-all-self-hosting/ansible-role-gotosocial
closed
Make account domain a setting
documentation enhancement
See: https://docs.gotosocial.org/en/latest/configuration/general/ Also add warning for it: > DO NOT change this after your server has already run once, or you will break things!
1.0
Make account domain a setting - See: https://docs.gotosocial.org/en/latest/configuration/general/ Also add warning for it: > DO NOT change this after your server has already run once, or you will break things!
non_process
make account domain a setting see also add warning for it do not change this after your server has already run once or you will break things
0
9,331
12,340,581,393
IssuesEvent
2020-05-14 20:12:15
googleapis/nodejs-billing
https://api.github.com/repos/googleapis/nodejs-billing
closed
GA Release of @google-cloud/billing
type: process
Package name: **@google-cloud/billing** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [x] Manual code has been reviewed by API producer - [x] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
GA Release of @google-cloud/billing - Package name: **@google-cloud/billing** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [x] Manual code has been reviewed by API producer - [x] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
ga release of google cloud billing package name google cloud billing current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
1,482
4,057,065,688
IssuesEvent
2016-05-24 20:48:10
pelias/pelias
https://api.github.com/repos/pelias/pelias
closed
Categorize POIs to facilitate result filtering [PL-PG06]
processed story
Categories will make reverse place geocoding a useful service. This could be done in conjunction with adding popularity scores/tags.
1.0
Categorize POIs to facilitate result filtering [PL-PG06] - Categories will make reverse place geocoding a useful service. This could be done in conjunction with adding popularity scores/tags.
process
categorize pois to facilitate result filtering categories will make reverse place geocoding a useful service this could be done in conjunction with adding popularity scores tags
1
3,661
6,694,648,150
IssuesEvent
2017-10-10 03:25:34
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
opened
Watering Assignment - Change Data Source
enhancement process workflow
Change the data source to pull directly from the @etrans tables.
1.0
Watering Assignment - Change Data Source - Change the data source to pull directly from the @etrans tables.
process
watering assignment change data source change the data source to pull directly from the etrans tables
1
251,355
27,151,181,920
IssuesEvent
2023-02-17 01:28:55
nidhi7598/linux-4.19.72
https://api.github.com/repos/nidhi7598/linux-4.19.72
reopened
CVE-2023-0266 (High) detected in linuxlinux-4.19.254
security vulnerability
## CVE-2023-0266 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ALSA: pcm: Move rwsem lock inside snd_ctl_elem_read to prevent UAF <p>Publish Date: 2023-01-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0266>CVE-2023-0266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0266">https://www.linuxkernelcves.com/cves/CVE-2023-0266</a></p> <p>Release Date: 2023-01-30</p> <p>Fix Resolution: v4.14.303,v4.19.270,v5.4.229,v5.10.163,v5.15.88,v6.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-0266 (High) detected in linuxlinux-4.19.254 - ## CVE-2023-0266 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ALSA: pcm: Move rwsem lock inside snd_ctl_elem_read to prevent UAF <p>Publish Date: 2023-01-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0266>CVE-2023-0266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0266">https://www.linuxkernelcves.com/cves/CVE-2023-0266</a></p> <p>Release Date: 2023-01-30</p> <p>Fix Resolution: v4.14.303,v4.19.270,v5.4.229,v5.10.163,v5.15.88,v6.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details alsa pcm move rwsem lock inside snd ctl elem read to prevent uaf publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required low user interaction none scope changed impact metrics confidentiality impact low integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,351
5,163,149,639
IssuesEvent
2017-01-17 04:58:38
inasafe/inasafe
https://api.github.com/repos/inasafe/inasafe
closed
Padang project consistently crashes QGIS
Bug Preprocessing
how to reproduce: - open AMCDRRbranch qgis_projects/padang.qgs - run, after a minute or so and taking up all Gb of ram and swap it segfaults during "calculating results" with this message: Debug: src/core/qgsmessagelog.cpp: 41: (logMessage) 2012-10-22T11:47:55 InaSAFE[0] gdalwarp -q -t_srs EPSG:4326 -r near -tr 0.000899 0.000899 -cutline /tmp/inasafe/22-10-2012/marco/work/extent_Mcjm3X.kml -crop_to_cutline -of GTiff "/home/marco/dev/qgis-plugins/inasafe_data/exposure/population_sumatra.tif" "/tmp/inasafe/22-10-2012/marco/work/clip_7EwcHH.tif" Debug: src/core/qgsmessagelog.cpp: 41: (logMessage) 2012-10-22T11:48:19 InaSAFE[0] calculate_impact called with: Layers: [<safe.storage.raster.Raster instance at 0x54d8680>, <safe.storage.raster.Raster instance at 0x54d86c8>] Function:<class 'safe.impact_functions.earthquake.itb_earthquake_fatality_model.ITBFatalityFunction'> Segmentation fault (core dumped)
1.0
Padang project consistently crashes QGIS - how to reproduce: - open AMCDRRbranch qgis_projects/padang.qgs - run, after a minute or so and taking up all Gb of ram and swap it segfaults during "calculating results" with this message: Debug: src/core/qgsmessagelog.cpp: 41: (logMessage) 2012-10-22T11:47:55 InaSAFE[0] gdalwarp -q -t_srs EPSG:4326 -r near -tr 0.000899 0.000899 -cutline /tmp/inasafe/22-10-2012/marco/work/extent_Mcjm3X.kml -crop_to_cutline -of GTiff "/home/marco/dev/qgis-plugins/inasafe_data/exposure/population_sumatra.tif" "/tmp/inasafe/22-10-2012/marco/work/clip_7EwcHH.tif" Debug: src/core/qgsmessagelog.cpp: 41: (logMessage) 2012-10-22T11:48:19 InaSAFE[0] calculate_impact called with: Layers: [<safe.storage.raster.Raster instance at 0x54d8680>, <safe.storage.raster.Raster instance at 0x54d86c8>] Function:<class 'safe.impact_functions.earthquake.itb_earthquake_fatality_model.ITBFatalityFunction'> Segmentation fault (core dumped)
process
padang project consistently crashes qgis how to reproduce open amcdrrbranch qgis projects padang qgs run after a minute or so and taking up all gb of ram and swap it segfaults during calculating results with this message debug src core qgsmessagelog cpp logmessage inasafe gdalwarp q t srs epsg r near tr cutline tmp inasafe marco work extent kml crop to cutline of gtiff home marco dev qgis plugins inasafe data exposure population sumatra tif tmp inasafe marco work clip tif debug src core qgsmessagelog cpp logmessage inasafe calculate impact called with layers function segmentation fault core dumped
1
497,721
14,383,507,606
IssuesEvent
2020-12-02 09:13:19
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
Potential race condition between processing starts and stream processor is paused
Priority: High Scope: broker Status: Needs Review Type: Maintenance
**Description** If processing already paused before a StreamProcessor is started, there is a small window in which processing can still happen. This is how StreamProcessor is paused during it is started. ``` openFuture.onComplete( (nothing, err) -> { if (err == null) { context.setStreamProcessor(streamProcessor); if (!context.shouldProcess()) { streamProcessor.pauseProcessing(); } ``` A solution will be to pass the `pause state` in the constructor of StreamProcessor as we did for the exporter in this PR https://github.com/zeebe-io/zeebe/pull/5903 . However, it still should complete reprocessing before it is actually sets its phase to paused.
1.0
Potential race condition between processing starts and stream processor is paused - **Description** If processing already paused before a StreamProcessor is started, there is a small window in which processing can still happen. This is how StreamProcessor is paused during it is started. ``` openFuture.onComplete( (nothing, err) -> { if (err == null) { context.setStreamProcessor(streamProcessor); if (!context.shouldProcess()) { streamProcessor.pauseProcessing(); } ``` A solution will be to pass the `pause state` in the constructor of StreamProcessor as we did for the exporter in this PR https://github.com/zeebe-io/zeebe/pull/5903 . However, it still should complete reprocessing before it is actually sets its phase to paused.
non_process
potential race condition between processing starts and stream processor is paused description if processing already paused before a streamprocessor is started there is a small window in which processing can still happen this is how streamprocessor is paused during it is started openfuture oncomplete nothing err if err null context setstreamprocessor streamprocessor if context shouldprocess streamprocessor pauseprocessing a solution will be to pass the pause state in the constructor of streamprocessor as we did for the exporter in this pr however it still should complete reprocessing before it is actually sets its phase to paused
0
3,092
6,107,774,777
IssuesEvent
2017-06-21 08:56:22
NBISweden/LocalEGA
https://api.github.com/repos/NBISweden/LocalEGA
opened
P3. 3. Requester Registration & Dataset Access
Process
1. Requester asks for access to a dataset through CEGA interface 2. DAC grants access to dataset -> standard route OR FUSE layer access Standard route Needs to be looked into more - could be provided by Alexander S’s code 3. CEGA messages LEGA instance to create an OUTBOX endpoint, and egress(?) of relevant dataset files from vault. 4. LEGA instance creates an OUTBOX for requester. 5. LEGA instance re-encrypts dataset files with requester-specific encryption credentials, and moves re-encrypted dataset files to OUTBOX. 6. LEGA instance messages requester (via CEGA or directly?) about endpoint address, protocol, and access/encryption credentials. 7. Requester downloads dataset files
1.0
P3. 3. Requester Registration & Dataset Access - 1. Requester asks for access to a dataset through CEGA interface 2. DAC grants access to dataset -> standard route OR FUSE layer access Standard route Needs to be looked into more - could be provided by Alexander S’s code 3. CEGA messages LEGA instance to create an OUTBOX endpoint, and egress(?) of relevant dataset files from vault. 4. LEGA instance creates an OUTBOX for requester. 5. LEGA instance re-encrypts dataset files with requester-specific encryption credentials, and moves re-encrypted dataset files to OUTBOX. 6. LEGA instance messages requester (via CEGA or directly?) about endpoint address, protocol, and access/encryption credentials. 7. Requester downloads dataset files
process
requester registration dataset access requester asks for access to a dataset through cega interface dac grants access to dataset standard route or fuse layer access standard route needs to be looked into more could be provided by alexander s’s code cega messages lega instance to create an outbox endpoint and egress of relevant dataset files from vault lega instance creates an outbox for requester lega instance re encrypts dataset files with requester specific encryption credentials and moves re encrypted dataset files to outbox lega instance messages requester via cega or directly about endpoint address protocol and access encryption credentials requester downloads dataset files
1
12,070
7,776,647,350
IssuesEvent
2018-06-05 08:44:31
ppy/osu
https://api.github.com/repos/ppy/osu
closed
Opening the player for multi while Circles is playing will reserve all memory
bug performance
Steps to reproduce: - Open osu!lazer - Change mode to osu!mania - Do NOT change the beatmap, let Circles play in the background - Navigate to Play > Multi > Match > Player Version: 2018.522.0 The reserved memory will start stacking up with a rate of ~300 MB/s, leading to severe lag once too much memory is reserved in an attempt to free it. Despite the fact that: - This has to do with the fact that any attempt to load an osu!mania beatmap not containing any objects won't throw an exception - Circles is not meant to be usable in any form when opening any screen I think it's useful to at least get rid of that bug. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/58678070-opening-the-player-for-multi-while-circles-is-playing-will-reserve-all-memory?utm_campaign=plugin&utm_content=tracker%2F4707269&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F4707269&utm_medium=issues&utm_source=github). </bountysource-plugin>
True
Opening the player for multi while Circles is playing will reserve all memory - Steps to reproduce: - Open osu!lazer - Change mode to osu!mania - Do NOT change the beatmap, let Circles play in the background - Navigate to Play > Multi > Match > Player Version: 2018.522.0 The reserved memory will start stacking up with a rate of ~300 MB/s, leading to severe lag once too much memory is reserved in an attempt to free it. Despite the fact that: - This has to do with the fact that any attempt to load an osu!mania beatmap not containing any objects won't throw an exception - Circles is not meant to be usable in any form when opening any screen I think it's useful to at least get rid of that bug. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/58678070-opening-the-player-for-multi-while-circles-is-playing-will-reserve-all-memory?utm_campaign=plugin&utm_content=tracker%2F4707269&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F4707269&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
opening the player for multi while circles is playing will reserve all memory steps to reproduce open osu lazer change mode to osu mania do not change the beatmap let circles play in the background navigate to play multi match player version the reserved memory will start stacking up with a rate of mb s leading to severe lag once too much memory is reserved in an attempt to free it despite the fact that this has to do with the fact that any attempt to load an osu mania beatmap not containing any objects won t throw an exception circles is not meant to be usable in any form when opening any screen i think it s useful to at least get rid of that bug want to back this issue we accept bounties via
0
17,242
23,019,361,183
IssuesEvent
2022-07-22 02:25:46
fmnas/fmnas-site
https://api.github.com/repos/fmnas/fmnas-site
closed
The serialized files don't get deleted after resending failed applications for some reason.
bug admin backend form processor x-small (<1h)
--- _This issue has been automatically created by [todo-actions](https://github.com/apps/todo-actions) based on a TODO comment found in [admin/resend_application.php:36](https://github.com/fmnas/fmnas-site/blob/main/admin/resend_application.php#L36). It will automatically be closed when the TODO comment is removed from the default branch (main)._
1.0
The serialized files don't get deleted after resending failed applications for some reason. - --- _This issue has been automatically created by [todo-actions](https://github.com/apps/todo-actions) based on a TODO comment found in [admin/resend_application.php:36](https://github.com/fmnas/fmnas-site/blob/main/admin/resend_application.php#L36). It will automatically be closed when the TODO comment is removed from the default branch (main)._
process
the serialized files don t get deleted after resending failed applications for some reason this issue has been automatically created by based on a todo comment found in it will automatically be closed when the todo comment is removed from the default branch main
1
731,035
25,199,173,699
IssuesEvent
2022-11-12 22:46:53
authelia/authelia
https://api.github.com/repos/authelia/authelia
reopened
Security of 2FA enrollment process may be improved.
type/enhancement priority/3/medium area/security
Hi, Let's assume the following scenario to illustrate the risk refered by this issue's title and explain how this may ruin the security benefit of 2FA. - An attacker manages to steal valid credentials. - The victim uses the same password to access his e-mail service. - The attacker uses these credentials access the victim's e-mails. - The attacker can enroll his own device as a 2nd factor to Authelia. - Victim's data protected by Authelia can be accessed and compromised by the attacker. As a result, the whole benefit of 2FA is defeated – assuming the user had the same password for both his Authelia account and e-mail service, and yes, password reuse is unfortunately very frequent. Therefore, I would advice the following behavior to lower such a risk. 1. A user wants to register a second factor. 2. An e-mail is sent to the admin – yes… another parameter in Authelia's configuration file –, which may approve or not this request. Such an e-mail may look like this: >`username` asked on `date` to enroll a new device as a second factor (`TOTP` | `U2F`) for its account on `login.example.com`. > >This request was sent from `IP` (`Country`, `Region`, `City` | unkown location *[if geo-location has failed]*). > >Do you approve this request? > >[a green “Yes” button] [a red “No” button] 3. If the admin approves this request, the user receives the e-mail he would receives in the current Authelia behavior, to actually register a second factor. Step 2. makes the aformentioned attack significantly more difficult, since the admin approval is required. The idea is that it may get in touch – physically, or by phone – with the user to make sure such a request was legitimate, in case it was unlikely to occur – weird IP location, user had already enrolled and has not reported loss or theft of its phone/YubiKey, etc. **Note**: In step 3., keep it as it currently is: it is very neat – you do not ever send any sensitive information through e-mail, evey link sent by Authelia is valid only one time and for a limited duration. 👍 If you decide to use IP location, using some kind of geo-location package, note that it wouldn't hurt to always include these details for every e-mail that is sent by Authelia: it helps to detect malicious activity.
1.0
Security of 2FA enrollment process may be improved. - Hi, Let's assume the following scenario to illustrate the risk refered by this issue's title and explain how this may ruin the security benefit of 2FA. - An attacker manages to steal valid credentials. - The victim uses the same password to access his e-mail service. - The attacker uses these credentials access the victim's e-mails. - The attacker can enroll his own device as a 2nd factor to Authelia. - Victim's data protected by Authelia can be accessed and compromised by the attacker. As a result, the whole benefit of 2FA is defeated – assuming the user had the same password for both his Authelia account and e-mail service, and yes, password reuse is unfortunately very frequent. Therefore, I would advice the following behavior to lower such a risk. 1. A user wants to register a second factor. 2. An e-mail is sent to the admin – yes… another parameter in Authelia's configuration file –, which may approve or not this request. Such an e-mail may look like this: >`username` asked on `date` to enroll a new device as a second factor (`TOTP` | `U2F`) for its account on `login.example.com`. > >This request was sent from `IP` (`Country`, `Region`, `City` | unkown location *[if geo-location has failed]*). > >Do you approve this request? > >[a green “Yes” button] [a red “No” button] 3. If the admin approves this request, the user receives the e-mail he would receives in the current Authelia behavior, to actually register a second factor. Step 2. makes the aformentioned attack significantly more difficult, since the admin approval is required. The idea is that it may get in touch – physically, or by phone – with the user to make sure such a request was legitimate, in case it was unlikely to occur – weird IP location, user had already enrolled and has not reported loss or theft of its phone/YubiKey, etc. **Note**: In step 3., keep it as it currently is: it is very neat – you do not ever send any sensitive information through e-mail, evey link sent by Authelia is valid only one time and for a limited duration. 👍 If you decide to use IP location, using some kind of geo-location package, note that it wouldn't hurt to always include these details for every e-mail that is sent by Authelia: it helps to detect malicious activity.
non_process
security of enrollment process may be improved hi let s assume the following scenario to illustrate the risk refered by this issue s title and explain how this may ruin the security benefit of an attacker manages to steal valid credentials the victim uses the same password to access his e mail service the attacker uses these credentials access the victim s e mails the attacker can enroll his own device as a factor to authelia victim s data protected by authelia can be accessed and compromised by the attacker as a result the whole benefit of is defeated – assuming the user had the same password for both his authelia account and e mail service and yes password reuse is unfortunately very frequent therefore i would advice the following behavior to lower such a risk a user wants to register a second factor an e mail is sent to the admin – yes… another parameter in authelia s configuration file – which may approve or not this request such an e mail may look like this username asked on date to enroll a new device as a second factor totp for its account on login example com this request was sent from ip country region city unkown location do you approve this request if the admin approves this request the user receives the e mail he would receives in the current authelia behavior to actually register a second factor step makes the aformentioned attack significantly more difficult since the admin approval is required the idea is that it may get in touch – physically or by phone – with the user to make sure such a request was legitimate in case it was unlikely to occur – weird ip location user had already enrolled and has not reported loss or theft of its phone yubikey etc note in step keep it as it currently is it is very neat – you do not ever send any sensitive information through e mail evey link sent by authelia is valid only one time and for a limited duration 👍 if you decide to use ip location using some kind of geo location package note that it wouldn t hurt to always include these details for every e mail that is sent by authelia it helps to detect malicious activity
0
281,006
21,315,323,840
IssuesEvent
2022-04-16 07:02:24
sophiien/pe
https://api.github.com/repos/sophiien/pe
opened
Confusing UG
severity.Medium type.DocumentationBug
**Describe the bug** Confusing UG for adding event **To Reproduce** Steps to reproduce the behavior: 1. Type "event 1 name/hello info/hello movies d/2022-12-12 t/13:00" **Expected behaviour** If I understand the UG correctly, I am supposed to add the first person to this event but the error shows that the person index provided is not valid. **Screenshots** ![Screenshot 2022-04-16 at 3.00.44 PM.png](https://raw.githubusercontent.com/sophiien/pe/main/files/6f5803f7-d7c2-4113-b9bf-0882997e5d72.png) Shows that person of index 1 does exist ![Screenshot 2022-04-16 at 3.01.24 PM.png](https://raw.githubusercontent.com/sophiien/pe/main/files/18fdfaef-5dc5-49dd-aac6-562842fc0a4e.png) **Desktop (please complete the following information):** OS: macOS 11.6 <!--session: 1650089584959-3c834ccd-971f-47e7-9307-a21d9af08010--> <!--Version: Web v3.4.2-->
1.0
Confusing UG - **Describe the bug** Confusing UG for adding event **To Reproduce** Steps to reproduce the behavior: 1. Type "event 1 name/hello info/hello movies d/2022-12-12 t/13:00" **Expected behaviour** If I understand the UG correctly, I am supposed to add the first person to this event but the error shows that the person index provided is not valid. **Screenshots** ![Screenshot 2022-04-16 at 3.00.44 PM.png](https://raw.githubusercontent.com/sophiien/pe/main/files/6f5803f7-d7c2-4113-b9bf-0882997e5d72.png) Shows that person of index 1 does exist ![Screenshot 2022-04-16 at 3.01.24 PM.png](https://raw.githubusercontent.com/sophiien/pe/main/files/18fdfaef-5dc5-49dd-aac6-562842fc0a4e.png) **Desktop (please complete the following information):** OS: macOS 11.6 <!--session: 1650089584959-3c834ccd-971f-47e7-9307-a21d9af08010--> <!--Version: Web v3.4.2-->
non_process
confusing ug describe the bug confusing ug for adding event to reproduce steps to reproduce the behavior type event name hello info hello movies d t expected behaviour if i understand the ug correctly i am supposed to add the first person to this event but the error shows that the person index provided is not valid screenshots shows that person of index does exist desktop please complete the following information os macos
0