Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
43,181
11,543,203,417
IssuesEvent
2020-02-18 09:10:19
contao/contao
https://api.github.com/repos/contao/contao
closed
addWizardClass breaks compatibility with older versions
defect
**Affected version(s)** 4.9.0-RC1 **Description** Any extension that uses any picker in a DCA which needed the `wizard` CSS-class will break the styling with 4.9 because of the new `addWizardClass` attribute. I've seen no way to be compatible with 4.4 and 4.9 without distinguishing between the versions within every `eval` of every field. ![image](https://user-images.githubusercontent.com/226890/74236063-a8601880-4cd0-11ea-87d0-9a0efaeb575e.png) > With wizard class from 4.4 ![image](https://user-images.githubusercontent.com/226890/74236194-01c84780-4cd1-11ea-9505-efd9de8fd6ef.png) > Without wizard class as of 4.9 Maybe the handling within https://github.com/contao/contao/blob/4.9/core-bundle/src/Resources/contao/classes/DataContainer.php#L520-L532 could be improved so 4.9 removes the wizard class if we explicitly say `'addWizardClass'=>false`. This way we'd have at least a chance to be backwards compatible.
1.0
addWizardClass breaks compatibility with older versions - **Affected version(s)** 4.9.0-RC1 **Description** Any extension that uses any picker in a DCA which needed the `wizard` CSS-class will break the styling with 4.9 because of the new `addWizardClass` attribute. I've seen no way to be compatible with 4.4 and 4.9 without distinguishing between the versions within every `eval` of every field. ![image](https://user-images.githubusercontent.com/226890/74236063-a8601880-4cd0-11ea-87d0-9a0efaeb575e.png) > With wizard class from 4.4 ![image](https://user-images.githubusercontent.com/226890/74236194-01c84780-4cd1-11ea-9505-efd9de8fd6ef.png) > Without wizard class as of 4.9 Maybe the handling within https://github.com/contao/contao/blob/4.9/core-bundle/src/Resources/contao/classes/DataContainer.php#L520-L532 could be improved so 4.9 removes the wizard class if we explicitly say `'addWizardClass'=>false`. This way we'd have at least a chance to be backwards compatible.
non_test
addwizardclass breaks compatibility with older versions affected version s description any extension that uses any picker in a dca which needed the wizard css class will break the styling with because of the new addwizardclass attribute i ve seen no way to be compatible with and without distinguishing between the versions within every eval of every field with wizard class from without wizard class as of maybe the handling within could be improved so removes the wizard class if we explicitly say addwizardclass false this way we d have at least a chance to be backwards compatible
0
143,310
11,542,494,128
IssuesEvent
2020-02-18 07:34:23
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
`--disable-sync` arg is displayed twice in an upgraded profile
QA/Test-Plan-Specified QA/Yes bug
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> `--disable-sync` arg is displayed twice in an upgraded profile ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 1.3.115 2. Upgrade to 1.3.116 3. Open chrome://version 4. `--disable-sync` argument is displayed twice in cmd line ## Actual result: <!--Please add screenshots if needed--> `--disable-sync` arg is displayed twice in an upgraded profile ## Expected result: `--disable-sync` arg should be displayed once in an upgraded profile ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.3.116 Chromium: 80.0.3987.106 (Official Build) (64-bit) -- | -- Revision | f68069574609230cf9b635cd784cfb1bf81bb53a-refs/branch-heads/3987@{#882} OS | Windows 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the dev channel? No - Can you reproduce this issue with the nightly channel? No ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @darkdh
1.0
`--disable-sync` arg is displayed twice in an upgraded profile - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> `--disable-sync` arg is displayed twice in an upgraded profile ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 1.3.115 2. Upgrade to 1.3.116 3. Open chrome://version 4. `--disable-sync` argument is displayed twice in cmd line ## Actual result: <!--Please add screenshots if needed--> `--disable-sync` arg is displayed twice in an upgraded profile ## Expected result: `--disable-sync` arg should be displayed once in an upgraded profile ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.3.116 Chromium: 80.0.3987.106 (Official Build) (64-bit) -- | -- Revision | f68069574609230cf9b635cd784cfb1bf81bb53a-refs/branch-heads/3987@{#882} OS | Windows 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the dev channel? No - Can you reproduce this issue with the nightly channel? No ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @darkdh
test
disable sync arg is displayed twice in an upgraded profile have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description disable sync arg is displayed twice in an upgraded profile steps to reproduce clean profile upgrade to open chrome version disable sync argument is displayed twice in cmd line actual result disable sync arg is displayed twice in an upgraded profile expected result disable sync arg should be displayed once in an upgraded profile reproduces how often always brave version brave version info brave chromium   official build   bit revision refs branch heads os windows  os version build version channel information can you reproduce this issue with the current release no can you reproduce this issue with the beta channel yes can you reproduce this issue with the dev channel no can you reproduce this issue with the nightly channel no other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome miscellaneous information cc brave legacy qa darkdh
1
158,089
24,784,861,310
IssuesEvent
2022-10-24 08:57:28
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
UX: Search history
design team/search-product design-estimated-M
After releasing the 'simple UI' in Sourcegraph 4.0, the[ team identified ](https://docs.google.com/document/d/1dqRPWJFu01_5gczjOiVq_DKLxNHvrWPIz5dNi6y--QY/edit#bookmark=id.bzwkh143ivh9)the need to re-introduce a way for users to access the search history, avoiding leaving the product without a solution for accessing search history for too long. On 'simple UI,' we shipped the [first experiment](https://github.com/sourcegraph/sourcegraph/issues/39817) of search history on the search field (detailed in the _previous work_ section below) and ran a[ user testing study ](https://docs.google.com/document/d/1e-R8mNQnZJTzc50Vlv1uvakbOsPFz0FwTnmYKk7whxs/edit) where we learned essential points on how to improve it. This issue it's a continuation, a subsequent iteration, of that effort. ## Define **`Problem statement:`** Forgetting is [essential](https://www.discovermagazine.com/mind/why-do-we-forget-things-it-may-make-the-mind-more-efficient) for the mental health of any intelligent organism; the biological goal of the brain's memory apparatus is not to preserve information but to help the brain make good decisions. A [study from Stack Overflow](https://stackoverflow.blog/2022/06/22/asked-and-answered-the-results-for-the-2022-developer-survey-are-here/) reveals that 62% of developers search the same repeatedly when trying to fix a problem. Based on this context, we assumed that without a search history, our product doesn't aid developers in remembering past (sometimes long and complex) search queries and eases the cognitive load of their repetitive workflow. Also, not having an easy-to-access and simple-to-use UX diminishes the value of the Search product in reducing friction in future code searches. **`Goal:`** Decision on UX to improve the search history on the search field. **`Design challenges:`** - How might we indicate to users that they're glancing at search history (not suggestions)? - How might we differentiate search suggestions and history in the same UI component? - How might we allow users to access their full history and filter it? - How might we show quick search history without users focusing on the search field or browsing the full history? ## Previous work ### Search home in simple UI #39817 **Key learning from past iteration and user testing:** - Users didn't understand that they were browsing search history, even with history icons next to the items; they thought they were looking at search suggestions or examples of queries. ## Design - [ ] [Low-fi design and discussions (Figma file).](https://www.figma.com/file/XAa0yGjuz9k95P0htDhfVE/UX%3A-Search-history?node-id=1%3A672) - [ ] [Hi-fi visual design (Figma file).](https://www.figma.com/file/XAa0yGjuz9k95P0htDhfVE/UX%3A-Search-history?node-id=53%3A464) - [ ] [Interactive prototype.](https://www.figma.com/proto/XAa0yGjuz9k95P0htDhfVE/UX%3A-Search-history?page-id=53%3A464&node-id=53%3A7940&viewport=-927%2C1084%2C0.87&scaling=min-zoom&starting-point-node-id=53%3A7940&hide-ui=1) - [ ] Designer review. ## Design process (Size: M) - Estimate: Two weeks - Weighting factors: - MEDIUM risk - MEDIUM clarity /cc @lguychard
2.0
UX: Search history - After releasing the 'simple UI' in Sourcegraph 4.0, the[ team identified ](https://docs.google.com/document/d/1dqRPWJFu01_5gczjOiVq_DKLxNHvrWPIz5dNi6y--QY/edit#bookmark=id.bzwkh143ivh9)the need to re-introduce a way for users to access the search history, avoiding leaving the product without a solution for accessing search history for too long. On 'simple UI,' we shipped the [first experiment](https://github.com/sourcegraph/sourcegraph/issues/39817) of search history on the search field (detailed in the _previous work_ section below) and ran a[ user testing study ](https://docs.google.com/document/d/1e-R8mNQnZJTzc50Vlv1uvakbOsPFz0FwTnmYKk7whxs/edit) where we learned essential points on how to improve it. This issue it's a continuation, a subsequent iteration, of that effort. ## Define **`Problem statement:`** Forgetting is [essential](https://www.discovermagazine.com/mind/why-do-we-forget-things-it-may-make-the-mind-more-efficient) for the mental health of any intelligent organism; the biological goal of the brain's memory apparatus is not to preserve information but to help the brain make good decisions. A [study from Stack Overflow](https://stackoverflow.blog/2022/06/22/asked-and-answered-the-results-for-the-2022-developer-survey-are-here/) reveals that 62% of developers search the same repeatedly when trying to fix a problem. Based on this context, we assumed that without a search history, our product doesn't aid developers in remembering past (sometimes long and complex) search queries and eases the cognitive load of their repetitive workflow. Also, not having an easy-to-access and simple-to-use UX diminishes the value of the Search product in reducing friction in future code searches. **`Goal:`** Decision on UX to improve the search history on the search field. **`Design challenges:`** - How might we indicate to users that they're glancing at search history (not suggestions)? - How might we differentiate search suggestions and history in the same UI component? - How might we allow users to access their full history and filter it? - How might we show quick search history without users focusing on the search field or browsing the full history? ## Previous work ### Search home in simple UI #39817 **Key learning from past iteration and user testing:** - Users didn't understand that they were browsing search history, even with history icons next to the items; they thought they were looking at search suggestions or examples of queries. ## Design - [ ] [Low-fi design and discussions (Figma file).](https://www.figma.com/file/XAa0yGjuz9k95P0htDhfVE/UX%3A-Search-history?node-id=1%3A672) - [ ] [Hi-fi visual design (Figma file).](https://www.figma.com/file/XAa0yGjuz9k95P0htDhfVE/UX%3A-Search-history?node-id=53%3A464) - [ ] [Interactive prototype.](https://www.figma.com/proto/XAa0yGjuz9k95P0htDhfVE/UX%3A-Search-history?page-id=53%3A464&node-id=53%3A7940&viewport=-927%2C1084%2C0.87&scaling=min-zoom&starting-point-node-id=53%3A7940&hide-ui=1) - [ ] Designer review. ## Design process (Size: M) - Estimate: Two weeks - Weighting factors: - MEDIUM risk - MEDIUM clarity /cc @lguychard
non_test
ux search history after releasing the simple ui in sourcegraph the need to re introduce a way for users to access the search history avoiding leaving the product without a solution for accessing search history for too long on simple ui we shipped the of search history on the search field detailed in the previous work section below and ran a where we learned essential points on how to improve it this issue it s a continuation a subsequent iteration of that effort define problem statement forgetting is for the mental health of any intelligent organism the biological goal of the brain s memory apparatus is not to preserve information but to help the brain make good decisions a reveals that of developers search the same repeatedly when trying to fix a problem based on this context we assumed that without a search history our product doesn t aid developers in remembering past sometimes long and complex search queries and eases the cognitive load of their repetitive workflow also not having an easy to access and simple to use ux diminishes the value of the search product in reducing friction in future code searches goal decision on ux to improve the search history on the search field design challenges how might we indicate to users that they re glancing at search history not suggestions how might we differentiate search suggestions and history in the same ui component how might we allow users to access their full history and filter it how might we show quick search history without users focusing on the search field or browsing the full history previous work search home in simple ui key learning from past iteration and user testing users didn t understand that they were browsing search history even with history icons next to the items they thought they were looking at search suggestions or examples of queries design designer review design process size m estimate two weeks weighting factors medium risk medium clarity cc lguychard
0
293,435
25,292,083,127
IssuesEvent
2022-11-17 01:37:13
woowacourse-teams/2022-sokdak
https://api.github.com/repos/woowacourse-teams/2022-sokdak
opened
[REFACTOR] JwtTokenProviderTest Context Caching
⚡️ back-end refactor test
### 설명 JwtTokenProviderTest가 Context Caching을 사용하고 있지 않음. ### 개선할 부분 - [ ] JwtTokenProviderTest가 Caching된 Context를 사용하도록 수정
1.0
[REFACTOR] JwtTokenProviderTest Context Caching - ### 설명 JwtTokenProviderTest가 Context Caching을 사용하고 있지 않음. ### 개선할 부분 - [ ] JwtTokenProviderTest가 Caching된 Context를 사용하도록 수정
test
jwttokenprovidertest context caching 설명 jwttokenprovidertest가 context caching을 사용하고 있지 않음 개선할 부분 jwttokenprovidertest가 caching된 context를 사용하도록 수정
1
28,699
5,532,494,318
IssuesEvent
2017-03-21 10:46:21
chartjs/Chart.js
https://api.github.com/repos/chartjs/Chart.js
closed
[BUG] Broken link to global configuration section in docs
Category: Bug Category: Documentation
There is a broken link in the documentation in this section: http://www.chartjs.org/docs/#doughnut-pie-chart-chart-options The link "global chart configuration options" links to: http://www.chartjs.org/docs/#global-chart-configuration It should link to: http://www.chartjs.org/docs/#chart-configuration-global-configuration At a second glance it looks like there are a few links pointing to both anchors, so they probably just need to be merged.
1.0
[BUG] Broken link to global configuration section in docs - There is a broken link in the documentation in this section: http://www.chartjs.org/docs/#doughnut-pie-chart-chart-options The link "global chart configuration options" links to: http://www.chartjs.org/docs/#global-chart-configuration It should link to: http://www.chartjs.org/docs/#chart-configuration-global-configuration At a second glance it looks like there are a few links pointing to both anchors, so they probably just need to be merged.
non_test
broken link to global configuration section in docs there is a broken link in the documentation in this section the link global chart configuration options links to it should link to at a second glance it looks like there are a few links pointing to both anchors so they probably just need to be merged
0
12,189
3,257,021,275
IssuesEvent
2015-10-20 16:04:19
akvo/akvo-caddisfly
https://api.github.com/repos/akvo/akvo-caddisfly
closed
2.4 Check illumination quality of captured image
Strip test
We need to check if the global illumination level of the image is acceptable, because we don't want to adapt it before doing the calibration. This means both checking the global illumination level,and the if the illumination is homogenous across the image. To do this, we need to determine a way to measure the global illumination level, This could be done as follows: 1) in a grayscale image, there should only be some limited percentage of pixels with the extreme values, so 0 and 255. We need to determine a good cut off value for this. If pixels have these extreme values, it means that they are either under exposed, or over exposed, and in both cases information is lost. 2) for the R, G, and B channels, this should also be checked, in order to check oversaturation 3) For the grayscale image, we should find some additional quality check to see if the histogram of values is as we want it. For this, we might want to look at the histogram for a few 'good' and 'bad' images, and see what we can learn from that. To determine if the illumination is homogenous accross the image, one option to use is to look at a few horizontal and vertical white lines of which we know the location. For each of these lines, we could say that the individual values of all the pixels on the line should not differ more than some small percentage (3% - 5%?) from the average value of all the pixels on the line. This will detect if there are local differences in the illumination.
1.0
2.4 Check illumination quality of captured image - We need to check if the global illumination level of the image is acceptable, because we don't want to adapt it before doing the calibration. This means both checking the global illumination level,and the if the illumination is homogenous across the image. To do this, we need to determine a way to measure the global illumination level, This could be done as follows: 1) in a grayscale image, there should only be some limited percentage of pixels with the extreme values, so 0 and 255. We need to determine a good cut off value for this. If pixels have these extreme values, it means that they are either under exposed, or over exposed, and in both cases information is lost. 2) for the R, G, and B channels, this should also be checked, in order to check oversaturation 3) For the grayscale image, we should find some additional quality check to see if the histogram of values is as we want it. For this, we might want to look at the histogram for a few 'good' and 'bad' images, and see what we can learn from that. To determine if the illumination is homogenous accross the image, one option to use is to look at a few horizontal and vertical white lines of which we know the location. For each of these lines, we could say that the individual values of all the pixels on the line should not differ more than some small percentage (3% - 5%?) from the average value of all the pixels on the line. This will detect if there are local differences in the illumination.
test
check illumination quality of captured image we need to check if the global illumination level of the image is acceptable because we don t want to adapt it before doing the calibration this means both checking the global illumination level and the if the illumination is homogenous across the image to do this we need to determine a way to measure the global illumination level this could be done as follows in a grayscale image there should only be some limited percentage of pixels with the extreme values so and we need to determine a good cut off value for this if pixels have these extreme values it means that they are either under exposed or over exposed and in both cases information is lost for the r g and b channels this should also be checked in order to check oversaturation for the grayscale image we should find some additional quality check to see if the histogram of values is as we want it for this we might want to look at the histogram for a few good and bad images and see what we can learn from that to determine if the illumination is homogenous accross the image one option to use is to look at a few horizontal and vertical white lines of which we know the location for each of these lines we could say that the individual values of all the pixels on the line should not differ more than some small percentage from the average value of all the pixels on the line this will detect if there are local differences in the illumination
1
405,005
27,500,808,709
IssuesEvent
2023-03-05 17:20:47
MultiMachineBuilder/MultiMachineBuilder
https://api.github.com/repos/MultiMachineBuilder/MultiMachineBuilder
closed
Nov 15: Document all items
Type: documentation Type: development
- [x] All block declarations - [x] All block classes - [x] All block functionality - [x] All item declarations - [x] All item classes - [x] All item functionality - [x] All game classes - [x] All game functionality - [x] All game menus
1.0
Nov 15: Document all items - - [x] All block declarations - [x] All block classes - [x] All block functionality - [x] All item declarations - [x] All item classes - [x] All item functionality - [x] All game classes - [x] All game functionality - [x] All game menus
non_test
nov document all items all block declarations all block classes all block functionality all item declarations all item classes all item functionality all game classes all game functionality all game menus
0
117,791
9,959,195,905
IssuesEvent
2019-07-06 03:46:30
TopicFriends/TopicFriends
https://api.github.com/repos/TopicFriends/TopicFriends
opened
Reduce copy-paste in TestCafe npm scripts
TestCafe
Eg on npm test:all. Extract common params into config file(s). Call one script from another.
1.0
Reduce copy-paste in TestCafe npm scripts - Eg on npm test:all. Extract common params into config file(s). Call one script from another.
test
reduce copy paste in testcafe npm scripts eg on npm test all extract common params into config file s call one script from another
1
65,090
6,948,510,385
IssuesEvent
2017-12-06 00:43:51
rancher/rke
https://api.github.com/repos/rancher/rke
closed
Check Docker version on host
bug resolved to-test
As we already talk to the Docker API, we can retrieve `ServerVersion` and compare to supported Docker versions set in rke.
1.0
Check Docker version on host - As we already talk to the Docker API, we can retrieve `ServerVersion` and compare to supported Docker versions set in rke.
test
check docker version on host as we already talk to the docker api we can retrieve serverversion and compare to supported docker versions set in rke
1
302,737
9,285,747,195
IssuesEvent
2019-03-21 08:22:52
SCM-NV/qmflows-namd
https://api.github.com/repos/SCM-NV/qmflows-namd
closed
Aux_fit basis for hybrid calculations
High Priority enhancement
The current input for hybrid calculations takes as aux_fit a basis that is too slow. The alternative is to take the cFIT basis for each atom, but there are several versions of different qualities. The idea is then to create a json file of this type: { "Sb": [6, 6, 6, 7, 9], } then the user in the hybrid input template chooses the quality. For example: aux_fit: verygood meaning that it takes the 4th value in the list. The scale should be: low, medium, good, verygood, excellent . In the input, in the &KIND section, it would look like this: &KIND Sb BASIS_SET AUX_FIT cFIT BASIS_SET DZVP-MOLOPT-SR-GTH-q5 GTH-PBE-q5 &END In addition to this, in the input template of the hybrid the following two keywords need to be added in &force_eval -> &DFT section BASIS_SET_FILE_NAME /home/v13/cp2k_basis/BASIS_ADMM BASIS_SET_FILE_NAME /home/v13/cp2k_basis/BASIS_ADMM_MOLOPT Also, in the hybrid template, the following keywords need to be REMOVED: poisson: periodic: "None" psolver: "MT"
1.0
Aux_fit basis for hybrid calculations - The current input for hybrid calculations takes as aux_fit a basis that is too slow. The alternative is to take the cFIT basis for each atom, but there are several versions of different qualities. The idea is then to create a json file of this type: { "Sb": [6, 6, 6, 7, 9], } then the user in the hybrid input template chooses the quality. For example: aux_fit: verygood meaning that it takes the 4th value in the list. The scale should be: low, medium, good, verygood, excellent . In the input, in the &KIND section, it would look like this: &KIND Sb BASIS_SET AUX_FIT cFIT BASIS_SET DZVP-MOLOPT-SR-GTH-q5 GTH-PBE-q5 &END In addition to this, in the input template of the hybrid the following two keywords need to be added in &force_eval -> &DFT section BASIS_SET_FILE_NAME /home/v13/cp2k_basis/BASIS_ADMM BASIS_SET_FILE_NAME /home/v13/cp2k_basis/BASIS_ADMM_MOLOPT Also, in the hybrid template, the following keywords need to be REMOVED: poisson: periodic: "None" psolver: "MT"
non_test
aux fit basis for hybrid calculations the current input for hybrid calculations takes as aux fit a basis that is too slow the alternative is to take the cfit basis for each atom but there are several versions of different qualities the idea is then to create a json file of this type sb then the user in the hybrid input template chooses the quality for example aux fit verygood meaning that it takes the value in the list the scale should be low medium good verygood excellent in the input in the kind section it would look like this kind sb basis set aux fit cfit basis set dzvp molopt sr gth gth pbe end in addition to this in the input template of the hybrid the following two keywords need to be added in force eval dft section basis set file name home basis basis admm basis set file name home basis basis admm molopt also in the hybrid template the following keywords need to be removed poisson periodic none psolver mt
0
720,019
24,776,196,258
IssuesEvent
2022-10-23 19:09:39
authelia/authelia
https://api.github.com/repos/authelia/authelia
closed
4.37 fails to connect to postgres database
type/bug status/in-progress priority/2/high
### Version v4.37.0 ### Deployment Method Docker ### Reverse Proxy Caddy ### Reverse Proxy Version 2.6.2 ### Description Authelia fails to connect to my postgres database on startup, when the exact same setup works with 4.36.9. It sounds related to https://github.com/authelia/authelia/issues/4245, except that's a mariadb. It is potentially related to https://github.com/authelia/authelia/issues/4247 as I use a secret for the storage password, but I'm not seeing the same error message. ### Reproduction All I have to do to reproduce is update to 4.37.0, but obviously there might be something specific to my setup causing the issue. I've provided both my `docker-compose.yml` and `configuration.yml`, hopefully that will help with finding a minimal reproduction. ### Expectations Minor version bumps of Authelia should work with the same setup. ### Logs ```shell time="2022-10-23T12:36:08Z" level=error msg="Failure running the storage provider startup check: error pinging database: failed to connect to `host=db user=authelia database=authelia`: dial error (dial tcp 172.27.0.3:4321: connect: connection refused)" stack="github.com/authelia/authelia/v4/internal/commands/root.go:277 doStartupChecks\ngithub.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun\ngithub.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute\ngithub.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC\ngithub.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute\ngithub.com/authelia/authelia/v4/cmd/authelia/main.go:10 main\nruntime/proc.go:250 main\nruntime/asm_amd64.s:1594 goexit" time="2022-10-23T12:36:08Z" level=fatal msg="The following providers had fatal failures during startup: storage" stack="github.com/authelia/authelia/v4/internal/commands/root.go:305 doStartupChecks\ngithub.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun\ngithub.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute\ngithub.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC\ngithub.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute\ngithub.com/authelia/authelia/v4/cmd/authelia/main.go:10 main\nruntime/proc.go:250 main\nruntime/asm_amd64.s:1594 goexit" Here's those stacktraces formatted: github.com/authelia/authelia/v4/internal/commands/root.go:277 doStartupChecks github.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun github.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute github.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC github.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute github.com/authelia/authelia/v4/cmd/authelia/main.go:10 main runtime/proc.go:250 main runtime/asm_amd64.s:1594 goexit github.com/authelia/authelia/v4/internal/commands/root.go:305 doStartupChecks github.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun github.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute github.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC github.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute github.com/authelia/authelia/v4/cmd/authelia/main.go:10 main runtime/proc.go:250 main runtime/asm_amd64.s:1594 goexit ``` ### Configuration ```yaml theme: auto # jwt_secret: secrets/jwt_secret totp: issuer: <my.domain> webauthn: disable: true authentication_backend: ldap: url: ldap://lldap base_dn: dc=lldap username_attribute: uid additional_users_dn: ou=people users_filter: (&(|({username_attribute}={input})({mail_attribute}={input}))(objectClass=person)) additional_groups_dn: ou=groups groups_filter: (member={dn}) group_name_attribute: cn mail_attribute: mail display_name_attribute: displayName ## The username and password of the admin user. user: uid=admin,ou=people,dc=lldap # password: secrets/lldap_admin_password access_control: rules: - domain: "*.<my.domain>" policy: two_factor subject: "user:<me>" session: name: <my.domain> domain: <my.domain> # secret: secrets/session_secret redis: host: authelia-redis port: 6379 storage: # encryption_key: secrets/encryption_key postgres: host: db database: authelia username: authelia # password: secrets/storage_password notifier: filesystem: filename: /config/notification-public.txt ``` ### Documentation My docker-compose.yml: ```yml version: "3.8" services: authelia: # Switching to 4.36 works image: authelia/authelia:4.37 container_name: authelia depends_on: - db - redis environment: - AUTHELIA_AUTHENTICATION_BACKEND_LDAP_PASSWORD_FILE=/run/secrets/lldap_admin_password - AUTHELIA_JWT_SECRET_FILE=/secrets/jwt_secret - AUTHELIA_SESSION_SECRET_FILE=/secrets/session_secret - AUTHELIA_STORAGE_ENCRYPTION_KEY_FILE=/secrets/storage_encryption_key - AUTHELIA_STORAGE_POSTGRES_PASSWORD_FILE=/secrets/storage_password volumes: - ./config:/config - ./secrets:/secrets secrets: - lldap_admin_password restart: unless-stopped db: image: postgres:14-alpine container_name: authelia-db environment: POSTGRES_PASSWORD_FILE: /run/secrets/storage_password POSTGRES_USER: authelia POSTGRES_DB: authelia secrets: - storage_password volumes: - ./db:/var/lib/postgresql/data restart: unless-stopped redis: image: redis:alpine container_name: authelia-redis volumes: - ./redis:/data restart: unless-stopped secrets: lldap_admin_password: file: ../lldap/secrets/admin_password ```
1.0
4.37 fails to connect to postgres database - ### Version v4.37.0 ### Deployment Method Docker ### Reverse Proxy Caddy ### Reverse Proxy Version 2.6.2 ### Description Authelia fails to connect to my postgres database on startup, when the exact same setup works with 4.36.9. It sounds related to https://github.com/authelia/authelia/issues/4245, except that's a mariadb. It is potentially related to https://github.com/authelia/authelia/issues/4247 as I use a secret for the storage password, but I'm not seeing the same error message. ### Reproduction All I have to do to reproduce is update to 4.37.0, but obviously there might be something specific to my setup causing the issue. I've provided both my `docker-compose.yml` and `configuration.yml`, hopefully that will help with finding a minimal reproduction. ### Expectations Minor version bumps of Authelia should work with the same setup. ### Logs ```shell time="2022-10-23T12:36:08Z" level=error msg="Failure running the storage provider startup check: error pinging database: failed to connect to `host=db user=authelia database=authelia`: dial error (dial tcp 172.27.0.3:4321: connect: connection refused)" stack="github.com/authelia/authelia/v4/internal/commands/root.go:277 doStartupChecks\ngithub.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun\ngithub.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute\ngithub.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC\ngithub.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute\ngithub.com/authelia/authelia/v4/cmd/authelia/main.go:10 main\nruntime/proc.go:250 main\nruntime/asm_amd64.s:1594 goexit" time="2022-10-23T12:36:08Z" level=fatal msg="The following providers had fatal failures during startup: storage" stack="github.com/authelia/authelia/v4/internal/commands/root.go:305 doStartupChecks\ngithub.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun\ngithub.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute\ngithub.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC\ngithub.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute\ngithub.com/authelia/authelia/v4/cmd/authelia/main.go:10 main\nruntime/proc.go:250 main\nruntime/asm_amd64.s:1594 goexit" Here's those stacktraces formatted: github.com/authelia/authelia/v4/internal/commands/root.go:277 doStartupChecks github.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun github.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute github.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC github.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute github.com/authelia/authelia/v4/cmd/authelia/main.go:10 main runtime/proc.go:250 main runtime/asm_amd64.s:1594 goexit github.com/authelia/authelia/v4/internal/commands/root.go:305 doStartupChecks github.com/authelia/authelia/v4/internal/commands/root.go:87 cmdRootRun github.com/spf13/cobra@v1.6.0/command.go:920 (*Command).execute github.com/spf13/cobra@v1.6.0/command.go:1040 (*Command).ExecuteC github.com/spf13/cobra@v1.6.0/command.go:968 (*Command).Execute github.com/authelia/authelia/v4/cmd/authelia/main.go:10 main runtime/proc.go:250 main runtime/asm_amd64.s:1594 goexit ``` ### Configuration ```yaml theme: auto # jwt_secret: secrets/jwt_secret totp: issuer: <my.domain> webauthn: disable: true authentication_backend: ldap: url: ldap://lldap base_dn: dc=lldap username_attribute: uid additional_users_dn: ou=people users_filter: (&(|({username_attribute}={input})({mail_attribute}={input}))(objectClass=person)) additional_groups_dn: ou=groups groups_filter: (member={dn}) group_name_attribute: cn mail_attribute: mail display_name_attribute: displayName ## The username and password of the admin user. user: uid=admin,ou=people,dc=lldap # password: secrets/lldap_admin_password access_control: rules: - domain: "*.<my.domain>" policy: two_factor subject: "user:<me>" session: name: <my.domain> domain: <my.domain> # secret: secrets/session_secret redis: host: authelia-redis port: 6379 storage: # encryption_key: secrets/encryption_key postgres: host: db database: authelia username: authelia # password: secrets/storage_password notifier: filesystem: filename: /config/notification-public.txt ``` ### Documentation My docker-compose.yml: ```yml version: "3.8" services: authelia: # Switching to 4.36 works image: authelia/authelia:4.37 container_name: authelia depends_on: - db - redis environment: - AUTHELIA_AUTHENTICATION_BACKEND_LDAP_PASSWORD_FILE=/run/secrets/lldap_admin_password - AUTHELIA_JWT_SECRET_FILE=/secrets/jwt_secret - AUTHELIA_SESSION_SECRET_FILE=/secrets/session_secret - AUTHELIA_STORAGE_ENCRYPTION_KEY_FILE=/secrets/storage_encryption_key - AUTHELIA_STORAGE_POSTGRES_PASSWORD_FILE=/secrets/storage_password volumes: - ./config:/config - ./secrets:/secrets secrets: - lldap_admin_password restart: unless-stopped db: image: postgres:14-alpine container_name: authelia-db environment: POSTGRES_PASSWORD_FILE: /run/secrets/storage_password POSTGRES_USER: authelia POSTGRES_DB: authelia secrets: - storage_password volumes: - ./db:/var/lib/postgresql/data restart: unless-stopped redis: image: redis:alpine container_name: authelia-redis volumes: - ./redis:/data restart: unless-stopped secrets: lldap_admin_password: file: ../lldap/secrets/admin_password ```
non_test
fails to connect to postgres database version deployment method docker reverse proxy caddy reverse proxy version description authelia fails to connect to my postgres database on startup when the exact same setup works with it sounds related to except that s a mariadb it is potentially related to as i use a secret for the storage password but i m not seeing the same error message reproduction all i have to do to reproduce is update to but obviously there might be something specific to my setup causing the issue i ve provided both my docker compose yml and configuration yml hopefully that will help with finding a minimal reproduction expectations minor version bumps of authelia should work with the same setup logs shell time level error msg failure running the storage provider startup check error pinging database failed to connect to host db user authelia database authelia dial error dial tcp connect connection refused stack github com authelia authelia internal commands root go dostartupchecks ngithub com authelia authelia internal commands root go cmdrootrun ngithub com cobra command go command execute ngithub com cobra command go command executec ngithub com cobra command go command execute ngithub com authelia authelia cmd authelia main go main nruntime proc go main nruntime asm s goexit time level fatal msg the following providers had fatal failures during startup storage stack github com authelia authelia internal commands root go dostartupchecks ngithub com authelia authelia internal commands root go cmdrootrun ngithub com cobra command go command execute ngithub com cobra command go command executec ngithub com cobra command go command execute ngithub com authelia authelia cmd authelia main go main nruntime proc go main nruntime asm s goexit here s those stacktraces formatted github com authelia authelia internal commands root go dostartupchecks github com authelia authelia internal commands root go cmdrootrun github com cobra command go command execute github com cobra command go command executec github com cobra command go command execute github com authelia authelia cmd authelia main go main runtime proc go main runtime asm s goexit github com authelia authelia internal commands root go dostartupchecks github com authelia authelia internal commands root go cmdrootrun github com cobra command go command execute github com cobra command go command executec github com cobra command go command execute github com authelia authelia cmd authelia main go main runtime proc go main runtime asm s goexit configuration yaml theme auto jwt secret secrets jwt secret totp issuer webauthn disable true authentication backend ldap url ldap lldap base dn dc lldap username attribute uid additional users dn ou people users filter username attribute input mail attribute input objectclass person additional groups dn ou groups groups filter member dn group name attribute cn mail attribute mail display name attribute displayname the username and password of the admin user user uid admin ou people dc lldap password secrets lldap admin password access control rules domain policy two factor subject user session name domain secret secrets session secret redis host authelia redis port storage encryption key secrets encryption key postgres host db database authelia username authelia password secrets storage password notifier filesystem filename config notification public txt documentation my docker compose yml yml version services authelia switching to works image authelia authelia container name authelia depends on db redis environment authelia authentication backend ldap password file run secrets lldap admin password authelia jwt secret file secrets jwt secret authelia session secret file secrets session secret authelia storage encryption key file secrets storage encryption key authelia storage postgres password file secrets storage password volumes config config secrets secrets secrets lldap admin password restart unless stopped db image postgres alpine container name authelia db environment postgres password file run secrets storage password postgres user authelia postgres db authelia secrets storage password volumes db var lib postgresql data restart unless stopped redis image redis alpine container name authelia redis volumes redis data restart unless stopped secrets lldap admin password file lldap secrets admin password
0
68,787
7,110,138,763
IssuesEvent
2018-01-17 09:36:28
Pleio/pleio_template
https://api.github.com/repos/Pleio/pleio_template
closed
Bewerken groep: aanpassen tekstvakken
kia prio 1 testen
Er zijn in het menu 'groep bewerken' nu twee opties voor tekstvakken: 1. Beschrijving 2. Introductie Het is nu onduidelijk wat van beide tekstvakken de functie en de uiteindelijke plaatsing is. - tekstvak 1 'beschrijving' is de beschrijving van de groep: wat is het onderwerp van groep, voor wie is het interessant, ed. Zichtbaar voor niet-leden m.a.w. een advertentie voor je groep. - tekstvak 2 'introductie' is voor leden van de groep. Wordt zichtbaar bovenin de homepagina van de groep. Hier kan de groepsbeheerder een memo plaatsen. Groepsleden kunnen de memo wegklikken, als de beheerder een nieuw memo heeft geschreven wordt deze weer zichtbaar. Afgesproken dat: - [ ] Tekstvak 2 'introductie' hoort niet thuis in bij 'groep bewerken'. Moet hier uit worden verwijderd (PRIO 1) - [x] Naam tekstvak 1 'beschrijving' aanpassen naar: 'beschrijving voor niet-leden' - [x] Tekstvak 2 'introductie' wordt een soort widget op de homepagina van de groep, altijd zichtbaar voor de beheerder. - [x] Naam tekstvak 2 'introductie' aanpassen naar: memo voor leden
1.0
Bewerken groep: aanpassen tekstvakken - Er zijn in het menu 'groep bewerken' nu twee opties voor tekstvakken: 1. Beschrijving 2. Introductie Het is nu onduidelijk wat van beide tekstvakken de functie en de uiteindelijke plaatsing is. - tekstvak 1 'beschrijving' is de beschrijving van de groep: wat is het onderwerp van groep, voor wie is het interessant, ed. Zichtbaar voor niet-leden m.a.w. een advertentie voor je groep. - tekstvak 2 'introductie' is voor leden van de groep. Wordt zichtbaar bovenin de homepagina van de groep. Hier kan de groepsbeheerder een memo plaatsen. Groepsleden kunnen de memo wegklikken, als de beheerder een nieuw memo heeft geschreven wordt deze weer zichtbaar. Afgesproken dat: - [ ] Tekstvak 2 'introductie' hoort niet thuis in bij 'groep bewerken'. Moet hier uit worden verwijderd (PRIO 1) - [x] Naam tekstvak 1 'beschrijving' aanpassen naar: 'beschrijving voor niet-leden' - [x] Tekstvak 2 'introductie' wordt een soort widget op de homepagina van de groep, altijd zichtbaar voor de beheerder. - [x] Naam tekstvak 2 'introductie' aanpassen naar: memo voor leden
test
bewerken groep aanpassen tekstvakken er zijn in het menu groep bewerken nu twee opties voor tekstvakken beschrijving introductie het is nu onduidelijk wat van beide tekstvakken de functie en de uiteindelijke plaatsing is tekstvak beschrijving is de beschrijving van de groep wat is het onderwerp van groep voor wie is het interessant ed zichtbaar voor niet leden m a w een advertentie voor je groep tekstvak introductie is voor leden van de groep wordt zichtbaar bovenin de homepagina van de groep hier kan de groepsbeheerder een memo plaatsen groepsleden kunnen de memo wegklikken als de beheerder een nieuw memo heeft geschreven wordt deze weer zichtbaar afgesproken dat tekstvak introductie hoort niet thuis in bij groep bewerken moet hier uit worden verwijderd prio naam tekstvak beschrijving aanpassen naar beschrijving voor niet leden tekstvak introductie wordt een soort widget op de homepagina van de groep altijd zichtbaar voor de beheerder naam tekstvak introductie aanpassen naar memo voor leden
1
512,883
14,911,326,852
IssuesEvent
2021-01-22 10:54:09
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
VIEW:ALL - option generates errors and restarts of the NTOPNG process
low-priority bug to be confirmed
I try to use the "VIEW:ALL" option in ntopng.conf with "-i=view:all"ntopng to use the existing interfaces under NTOPNG as one view. But the ntopng process then goes into a reboot-loop. NTOPNG version: v.4.3.210119 [Enterprise/Professional build] Built on: Debian GNU/Linux 10 (buster) ntopng.conf: -G=/var/run/ntopng.pid -X=500000 -i="tcp://127.0.0.1:5556" -i="tcp://127.0.0.1:5557" -i=view:all -w=3000 --user=ntopng --dns-mode=1 root@CDC-S-NTOP01:/etc/ntopng# nano ntopng.conf root@CDC-S-NTOP01:/etc/ntopng# systemctl restart ntopng.service root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: active (running) since Tue 2021-01-19 14:30:37 CET; 3s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Main PID: 82574 (ntopng) Tasks: 51 (limit: 19660) Memory: 189.6M CGroup: /system.slice/ntopng.service └─82574 /usr/bin/ntopng /run/ntopng.conf Jan 19 14:30:39 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:39 [NtopPro.cpp:729] [LICENSE] Validity: Until Fri Jan 22 13:43:03 2021 Jan 19 14:30:39 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:39 [PeriodicActivities.cpp:107] Started periodic activities loop... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:42] Processing startup.lua: please hold on... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:158] [lists_utils.lua:753] Refreshing category lists... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:158] [lists_utils.lua:691] Category Lists (694 hosts, 4022 IPs, 98 JA3) loaded in 0 sec Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:232] Completed startup.lua Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [PeriodicActivities.cpp:168] Each periodic activity script will use 4 threads Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [NetworkInterface.cpp:2799] Started packet polling on interface tcp://127.0.0.1:5556 [id: 5]... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [NetworkInterface.cpp:2799] Started packet polling on interface tcp://127.0.0.1:5557 [id: 7]... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [NetworkInterface.cpp:2799] Started packet polling on interface view:all [id: 11]... root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: signal) since Tue 2021-01-19 14:30:41 CET; 2s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Process: 82574 ExecStart=/usr/bin/ntopng /run/ntopng.conf (code=killed, signal=SEGV) Process: 82699 ExecStopPost=/bin/rm -rf /run/ntopng.conf /run/ntopng.conf.raw /run/ntopng.pid (code=exited, status=0/SUCCESS) Main PID: 82574 (code=killed, signal=SEGV) root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: signal) since Tue 2021-01-19 14:30:41 CET; 3s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Process: 82574 ExecStart=/usr/bin/ntopng /run/ntopng.conf (code=killed, signal=SEGV) Process: 82699 ExecStopPost=/bin/rm -rf /run/ntopng.conf /run/ntopng.conf.raw /run/ntopng.pid (code=exited, status=0/SUCCESS) Main PID: 82574 (code=killed, signal=SEGV) root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: signal) since Tue 2021-01-19 14:30:41 CET; 4s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Process: 82574 ExecStart=/usr/bin/ntopng /run/ntopng.conf (code=killed, signal=SEGV) Process: 82699 ExecStopPost=/bin/rm -rf /run/ntopng.conf /run/ntopng.conf.raw /run/ntopng.pid (code=exited, status=0/SUCCESS) Main PID: 82574 (code=killed, signal=SEGV) root@CDC-S-NTOP01:/etc/ntopng#
1.0
VIEW:ALL - option generates errors and restarts of the NTOPNG process - I try to use the "VIEW:ALL" option in ntopng.conf with "-i=view:all"ntopng to use the existing interfaces under NTOPNG as one view. But the ntopng process then goes into a reboot-loop. NTOPNG version: v.4.3.210119 [Enterprise/Professional build] Built on: Debian GNU/Linux 10 (buster) ntopng.conf: -G=/var/run/ntopng.pid -X=500000 -i="tcp://127.0.0.1:5556" -i="tcp://127.0.0.1:5557" -i=view:all -w=3000 --user=ntopng --dns-mode=1 root@CDC-S-NTOP01:/etc/ntopng# nano ntopng.conf root@CDC-S-NTOP01:/etc/ntopng# systemctl restart ntopng.service root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: active (running) since Tue 2021-01-19 14:30:37 CET; 3s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Main PID: 82574 (ntopng) Tasks: 51 (limit: 19660) Memory: 189.6M CGroup: /system.slice/ntopng.service └─82574 /usr/bin/ntopng /run/ntopng.conf Jan 19 14:30:39 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:39 [NtopPro.cpp:729] [LICENSE] Validity: Until Fri Jan 22 13:43:03 2021 Jan 19 14:30:39 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:39 [PeriodicActivities.cpp:107] Started periodic activities loop... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:42] Processing startup.lua: please hold on... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:158] [lists_utils.lua:753] Refreshing category lists... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:158] [lists_utils.lua:691] Category Lists (694 hosts, 4022 IPs, 98 JA3) loaded in 0 sec Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [startup.lua:232] Completed startup.lua Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [PeriodicActivities.cpp:168] Each periodic activity script will use 4 threads Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [NetworkInterface.cpp:2799] Started packet polling on interface tcp://127.0.0.1:5556 [id: 5]... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [NetworkInterface.cpp:2799] Started packet polling on interface tcp://127.0.0.1:5557 [id: 7]... Jan 19 14:30:40 CDC-S-NTOP01 ntopng[82574]: 19/Jan/2021 14:30:40 [NetworkInterface.cpp:2799] Started packet polling on interface view:all [id: 11]... root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: signal) since Tue 2021-01-19 14:30:41 CET; 2s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Process: 82574 ExecStart=/usr/bin/ntopng /run/ntopng.conf (code=killed, signal=SEGV) Process: 82699 ExecStopPost=/bin/rm -rf /run/ntopng.conf /run/ntopng.conf.raw /run/ntopng.pid (code=exited, status=0/SUCCESS) Main PID: 82574 (code=killed, signal=SEGV) root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: signal) since Tue 2021-01-19 14:30:41 CET; 3s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Process: 82574 ExecStart=/usr/bin/ntopng /run/ntopng.conf (code=killed, signal=SEGV) Process: 82699 ExecStopPost=/bin/rm -rf /run/ntopng.conf /run/ntopng.conf.raw /run/ntopng.pid (code=exited, status=0/SUCCESS) Main PID: 82574 (code=killed, signal=SEGV) root@CDC-S-NTOP01:/etc/ntopng# systemctl status ntopng.service ● ntopng.service - ntopng high-speed web-based traffic monitoring and analysis tool Loaded: loaded (/etc/systemd/system/ntopng.service; enabled; vendor preset: enabled) Active: activating (auto-restart) (Result: signal) since Tue 2021-01-19 14:30:41 CET; 4s ago Process: 82555 ExecStartPre=/bin/sh -c /usr/bin/ntopng-utils-manage-config -a check-restore && /usr/bin/ntopng-utils-manage-config -a restore || true (code=exited, status=0/SUCCESS) Process: 82568 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf > /run/ntopng.conf.raw (code=exited, status=0/SUCCESS) Process: 82570 ExecStartPre=/bin/sh -c /bin/cat /etc/ntopng/ntopng.conf.d/*.conf >> /run/ntopng.conf.raw 2>/dev/null || true (code=exited, status=0/SUCCESS) Process: 82572 ExecStartPre=/bin/sh -c /bin/sed "/^[ ]*-e.*$\|^[ ]*-G.*\|^[ ]*--daemon.*\|[ ]*--pid.*/s/^/#/" /run/ntopng.conf.raw > /run/ntopng.conf (code=exited, status=0/SUCCESS) Process: 82574 ExecStart=/usr/bin/ntopng /run/ntopng.conf (code=killed, signal=SEGV) Process: 82699 ExecStopPost=/bin/rm -rf /run/ntopng.conf /run/ntopng.conf.raw /run/ntopng.pid (code=exited, status=0/SUCCESS) Main PID: 82574 (code=killed, signal=SEGV) root@CDC-S-NTOP01:/etc/ntopng#
non_test
view all option generates errors and restarts of the ntopng process i try to use the view all option in ntopng conf with i view all ntopng to use the existing interfaces under ntopng as one view but the ntopng process then goes into a reboot loop ntopng version v built on debian gnu linux buster ntopng conf g var run ntopng pid x i tcp i tcp i view all w user ntopng dns mode root cdc s etc ntopng nano ntopng conf root cdc s etc ntopng systemctl restart ntopng service root cdc s etc ntopng systemctl status ntopng service ● ntopng service ntopng high speed web based traffic monitoring and analysis tool loaded loaded etc systemd system ntopng service enabled vendor preset enabled active active running since tue cet ago process execstartpre bin sh c usr bin ntopng utils manage config a check restore usr bin ntopng utils manage config a restore true code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf run ntopng conf raw code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf d conf run ntopng conf raw dev null true code exited status success process execstartpre bin sh c bin sed e g daemon pid s run ntopng conf raw run ntopng conf code exited status success main pid ntopng tasks limit memory cgroup system slice ntopng service └─ usr bin ntopng run ntopng conf jan cdc s ntopng jan validity until fri jan jan cdc s ntopng jan started periodic activities loop jan cdc s ntopng jan processing startup lua please hold on jan cdc s ntopng jan refreshing category lists jan cdc s ntopng jan category lists hosts ips loaded in sec jan cdc s ntopng jan completed startup lua jan cdc s ntopng jan each periodic activity script will use threads jan cdc s ntopng jan started packet polling on interface tcp jan cdc s ntopng jan started packet polling on interface tcp jan cdc s ntopng jan started packet polling on interface view all root cdc s etc ntopng systemctl status ntopng service ● ntopng service ntopng high speed web based traffic monitoring and analysis tool loaded loaded etc systemd system ntopng service enabled vendor preset enabled active activating auto restart result signal since tue cet ago process execstartpre bin sh c usr bin ntopng utils manage config a check restore usr bin ntopng utils manage config a restore true code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf run ntopng conf raw code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf d conf run ntopng conf raw dev null true code exited status success process execstartpre bin sh c bin sed e g daemon pid s run ntopng conf raw run ntopng conf code exited status success process execstart usr bin ntopng run ntopng conf code killed signal segv process execstoppost bin rm rf run ntopng conf run ntopng conf raw run ntopng pid code exited status success main pid code killed signal segv root cdc s etc ntopng systemctl status ntopng service ● ntopng service ntopng high speed web based traffic monitoring and analysis tool loaded loaded etc systemd system ntopng service enabled vendor preset enabled active activating auto restart result signal since tue cet ago process execstartpre bin sh c usr bin ntopng utils manage config a check restore usr bin ntopng utils manage config a restore true code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf run ntopng conf raw code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf d conf run ntopng conf raw dev null true code exited status success process execstartpre bin sh c bin sed e g daemon pid s run ntopng conf raw run ntopng conf code exited status success process execstart usr bin ntopng run ntopng conf code killed signal segv process execstoppost bin rm rf run ntopng conf run ntopng conf raw run ntopng pid code exited status success main pid code killed signal segv root cdc s etc ntopng systemctl status ntopng service ● ntopng service ntopng high speed web based traffic monitoring and analysis tool loaded loaded etc systemd system ntopng service enabled vendor preset enabled active activating auto restart result signal since tue cet ago process execstartpre bin sh c usr bin ntopng utils manage config a check restore usr bin ntopng utils manage config a restore true code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf run ntopng conf raw code exited status success process execstartpre bin sh c bin cat etc ntopng ntopng conf d conf run ntopng conf raw dev null true code exited status success process execstartpre bin sh c bin sed e g daemon pid s run ntopng conf raw run ntopng conf code exited status success process execstart usr bin ntopng run ntopng conf code killed signal segv process execstoppost bin rm rf run ntopng conf run ntopng conf raw run ntopng pid code exited status success main pid code killed signal segv root cdc s etc ntopng
0
237,570
19,657,936,023
IssuesEvent
2022-01-10 14:22:59
astrolabsoftware/fink-science
https://api.github.com/repos/astrolabsoftware/fink-science
opened
[CI] Test with higher Python versions
test ci
We should also test the code with Python 3.7 & 3.8 at the very least.
1.0
[CI] Test with higher Python versions - We should also test the code with Python 3.7 & 3.8 at the very least.
test
test with higher python versions we should also test the code with python at the very least
1
145,054
19,319,022,477
IssuesEvent
2021-12-14 01:50:53
peterwkc85/selenium-jupiter
https://api.github.com/repos/peterwkc85/selenium-jupiter
opened
CVE-2020-35490 (High) detected in jackson-databind-2.9.8.jar
security vulnerability
## CVE-2020-35490 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /selenium-jupiter/build.gradle</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - docker-client-8.15.2.jar (Root Library) - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2020-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35490>CVE-2020-35490</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p> <p>Release Date: 2020-12-17</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-35490 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-35490 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /selenium-jupiter/build.gradle</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - docker-client-8.15.2.jar (Root Library) - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2020-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35490>CVE-2020-35490</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p> <p>Release Date: 2020-12-17</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file selenium jupiter build gradle path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy docker client jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
254,241
21,773,208,464
IssuesEvent
2022-05-13 11:12:19
ansible/ansible-builder
https://api.github.com/repos/ansible/ansible-builder
closed
Add ability to just generate contents of build folder
enhancement state:needs_test
It would be helpful to allow of just running the generate build folder command, rather then a full build. We have a use case downstream, where we want to generate the latest _build folder contents, but build using a different container build system. Today we run ansible-builder, but ctrl-C when the 2nd stage builds start.
1.0
Add ability to just generate contents of build folder - It would be helpful to allow of just running the generate build folder command, rather then a full build. We have a use case downstream, where we want to generate the latest _build folder contents, but build using a different container build system. Today we run ansible-builder, but ctrl-C when the 2nd stage builds start.
test
add ability to just generate contents of build folder it would be helpful to allow of just running the generate build folder command rather then a full build we have a use case downstream where we want to generate the latest build folder contents but build using a different container build system today we run ansible builder but ctrl c when the stage builds start
1
200
2,659,907,078
IssuesEvent
2015-03-19 00:26:22
isislovecruft/python-gnupg
https://api.github.com/repos/isislovecruft/python-gnupg
closed
python<=2.6 compatibility is broken by latest commit in master
bug compatibility patch
hi. master branch breaks py2 compatiblity using `sys.version_info.major` to check it is indeed py2. on py2 this is just a tuple. on py3 this is an object with attributes, one of which is referred to by line 55 in file `_utils.py`. My patch simply replaces the code with py2 safe code. I looked at "six" module in hope to find some kind of twofold compatibility, i.e. to write in py3 syntax, yet to still be compatible with py2... well it seems with sys there's a problem there.
True
python<=2.6 compatibility is broken by latest commit in master - hi. master branch breaks py2 compatiblity using `sys.version_info.major` to check it is indeed py2. on py2 this is just a tuple. on py3 this is an object with attributes, one of which is referred to by line 55 in file `_utils.py`. My patch simply replaces the code with py2 safe code. I looked at "six" module in hope to find some kind of twofold compatibility, i.e. to write in py3 syntax, yet to still be compatible with py2... well it seems with sys there's a problem there.
non_test
python compatibility is broken by latest commit in master hi master branch breaks compatiblity using sys version info major to check it is indeed on this is just a tuple on this is an object with attributes one of which is referred to by line in file utils py my patch simply replaces the code with safe code i looked at six module in hope to find some kind of twofold compatibility i e to write in syntax yet to still be compatible with well it seems with sys there s a problem there
0
128,395
27,252,089,589
IssuesEvent
2023-02-22 08:55:02
carstenbauer/MPITape.jl
https://api.github.com/repos/carstenbauer/MPITape.jl
opened
Rethink file structure
codework
`api.jl` `printing.jl` `fileio.jl`: all kind of arbitrary. Rethink this :)
1.0
Rethink file structure - `api.jl` `printing.jl` `fileio.jl`: all kind of arbitrary. Rethink this :)
non_test
rethink file structure api jl printing jl fileio jl all kind of arbitrary rethink this
0
616,318
19,299,056,546
IssuesEvent
2021-12-13 01:24:00
cdk8s-team/cdk8s-core
https://api.github.com/repos/cdk8s-team/cdk8s-core
closed
Yaml.stringify() and Yaml.parse()
feature-request priority/p2 effort/small stale
In addition to `load()` and `save()` which work with files, we should also expose in-memory versions. ### Description of the feature or enhancement: I would love to have some YAML utilities built-in to cdk8s. ### Use Case: <!-- why do you need this feature? --> While working with k8s, we are working with YAML in one way or another. Having some YAML utilities would enable users to do things like this one, without having to install external deps: ```typescript const datasources = new ConfigMap(this, 'grafana-datasources', { data: { "loki.yaml": yaml.safeDump({ apiVersion: 1, datasources: [{ name: 'Loki', type: 'loki', access: 'proxy', url: `${loki.releaseName}:3100`, jsonData: { maxLines: 1000 } }] }) } }); ``` ### Proposed Solution: <!-- Please include prototype/workaround/sketch/reference implementation: --> ### Other: <!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, gitter, etc --> This was proposed by @eladb on Slack. * [ ] :wave: I may be able to implement this feature request -> I don't have enough bandwidth right now :( * [ ] :warning: This feature might incur a breaking change --- This is a :rocket: Feature Request
1.0
Yaml.stringify() and Yaml.parse() - In addition to `load()` and `save()` which work with files, we should also expose in-memory versions. ### Description of the feature or enhancement: I would love to have some YAML utilities built-in to cdk8s. ### Use Case: <!-- why do you need this feature? --> While working with k8s, we are working with YAML in one way or another. Having some YAML utilities would enable users to do things like this one, without having to install external deps: ```typescript const datasources = new ConfigMap(this, 'grafana-datasources', { data: { "loki.yaml": yaml.safeDump({ apiVersion: 1, datasources: [{ name: 'Loki', type: 'loki', access: 'proxy', url: `${loki.releaseName}:3100`, jsonData: { maxLines: 1000 } }] }) } }); ``` ### Proposed Solution: <!-- Please include prototype/workaround/sketch/reference implementation: --> ### Other: <!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, gitter, etc --> This was proposed by @eladb on Slack. * [ ] :wave: I may be able to implement this feature request -> I don't have enough bandwidth right now :( * [ ] :warning: This feature might incur a breaking change --- This is a :rocket: Feature Request
non_test
yaml stringify and yaml parse in addition to load and save which work with files we should also expose in memory versions description of the feature or enhancement i would love to have some yaml utilities built in to use case while working with we are working with yaml in one way or another having some yaml utilities would enable users to do things like this one without having to install external deps typescript const datasources new configmap this grafana datasources data loki yaml yaml safedump apiversion datasources name loki type loki access proxy url loki releasename jsondata maxlines proposed solution other e g detailed explanation stacktraces related issues suggestions on how to fix links for us to have context eg associated pull request stackoverflow gitter etc this was proposed by eladb on slack wave i may be able to implement this feature request i don t have enough bandwidth right now warning this feature might incur a breaking change this is a rocket feature request
0
105
2,524,418,104
IssuesEvent
2015-01-20 17:36:24
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
Radio AMQP sender should use shorter reconnect timeout
infrastructure
Also make the connect timeout configurable, and optionally add a network recovery. refer to https://www.rabbitmq.com/api-guide.html#recovery for details on the latter.
1.0
Radio AMQP sender should use shorter reconnect timeout - Also make the connect timeout configurable, and optionally add a network recovery. refer to https://www.rabbitmq.com/api-guide.html#recovery for details on the latter.
non_test
radio amqp sender should use shorter reconnect timeout also make the connect timeout configurable and optionally add a network recovery refer to for details on the latter
0
264,177
23,099,930,992
IssuesEvent
2022-07-27 00:53:15
microsoft/vscode-jupyter
https://api.github.com/repos/microsoft/vscode-jupyter
closed
TestPlan: DataFrame viewer on the web
testplan-item
<!-- Header Section. --> Testing the DataFrame viewer on the web. Refs: https://github.com/microsoft/vscode-jupyter/issues/9665 - [x] web @rebornix Complexity: 3 Authors: @sadasant [Create Issue](https://github.com/microsoft/vscode-jupyter/issues/new?body=Testing+%2310915%0A%0A&assignees=sadasant) --- <!-- Body Section. --> ### Requirements - Install prerelease jupyter extension into http://insiders.vscode.dev - Have a python environment that you can run jupyter from. (https://jupyter.org/install) - Be familiar with jupyter notebooks ### Set up 1. Go to a command prompt where you've installed jupyter into a python environment 2. Switch to a folder you're using for testing 3. Run jupyter with the following command `jupyter notebook --no-browser --NotebookApp.allow_origin=*` ### Test steps 1. Open insiders.vscode.dev and create a new `.ipynb` file. 2. Connet to the jupyter server: 1. Run the 'Jupyter: Specify jupyter server for connections' command. 2. Pick 'Existing'. 3. Paste the URL from the jupyter notebook server you started in step 3 of the setup. 3. In one of the Notebook cells, enter one or more array variables of any length. 4. Run the cell in any way. For example: by clicking the ▶️ button at the left of the cell. 1. It might ask which kernel to use. Pick the remote kernel. 5. Run the command `Jupyter: Focus on Variables View`. 6. See that the variables in the Variable View matched the expectations. 7. In the Variables View, hover over any of the array variables. An icon will appear at the left-most column of the table. Click on that icon. 8. An expanded view of that variable should replace the notebook in the editor view. 9. Try that again with any other variable. https://user-images.githubusercontent.com/417016/180869130-68b352ba-a334-4aab-8f5b-038f2f8b221d.mov
1.0
TestPlan: DataFrame viewer on the web - <!-- Header Section. --> Testing the DataFrame viewer on the web. Refs: https://github.com/microsoft/vscode-jupyter/issues/9665 - [x] web @rebornix Complexity: 3 Authors: @sadasant [Create Issue](https://github.com/microsoft/vscode-jupyter/issues/new?body=Testing+%2310915%0A%0A&assignees=sadasant) --- <!-- Body Section. --> ### Requirements - Install prerelease jupyter extension into http://insiders.vscode.dev - Have a python environment that you can run jupyter from. (https://jupyter.org/install) - Be familiar with jupyter notebooks ### Set up 1. Go to a command prompt where you've installed jupyter into a python environment 2. Switch to a folder you're using for testing 3. Run jupyter with the following command `jupyter notebook --no-browser --NotebookApp.allow_origin=*` ### Test steps 1. Open insiders.vscode.dev and create a new `.ipynb` file. 2. Connet to the jupyter server: 1. Run the 'Jupyter: Specify jupyter server for connections' command. 2. Pick 'Existing'. 3. Paste the URL from the jupyter notebook server you started in step 3 of the setup. 3. In one of the Notebook cells, enter one or more array variables of any length. 4. Run the cell in any way. For example: by clicking the ▶️ button at the left of the cell. 1. It might ask which kernel to use. Pick the remote kernel. 5. Run the command `Jupyter: Focus on Variables View`. 6. See that the variables in the Variable View matched the expectations. 7. In the Variables View, hover over any of the array variables. An icon will appear at the left-most column of the table. Click on that icon. 8. An expanded view of that variable should replace the notebook in the editor view. 9. Try that again with any other variable. https://user-images.githubusercontent.com/417016/180869130-68b352ba-a334-4aab-8f5b-038f2f8b221d.mov
test
testplan dataframe viewer on the web testing the dataframe viewer on the web refs web rebornix complexity authors sadasant requirements install prerelease jupyter extension into have a python environment that you can run jupyter from be familiar with jupyter notebooks set up go to a command prompt where you ve installed jupyter into a python environment switch to a folder you re using for testing run jupyter with the following command jupyter notebook no browser notebookapp allow origin test steps open insiders vscode dev and create a new ipynb file connet to the jupyter server run the jupyter specify jupyter server for connections command pick existing paste the url from the jupyter notebook server you started in step of the setup in one of the notebook cells enter one or more array variables of any length run the cell in any way for example by clicking the ▶️ button at the left of the cell it might ask which kernel to use pick the remote kernel run the command jupyter focus on variables view see that the variables in the variable view matched the expectations in the variables view hover over any of the array variables an icon will appear at the left most column of the table click on that icon an expanded view of that variable should replace the notebook in the editor view try that again with any other variable
1
592,928
17,934,256,281
IssuesEvent
2021-09-10 13:29:05
medic/cht-core
https://api.github.com/repos/medic/cht-core
opened
Login page is not cached correctly
Type: Bug Regression Priority: 2 - Medium
**Describe the bug** If a user logs off while being disconnected from the internet, they won't be able to access the login page. **To Reproduce** Steps to reproduce the behavior: 1. Login as an offline user. 2. Disconnect from the internet. 3. Logout. 4. You're seeing your browser's "No internet" page instead of the cached login page. **Expected behavior** Login page should be cached: https://github.com/medic/cht-core/pull/5557 **Environment** - Instance: (eg: alpha.dev.medicmobile.org, etc) - Browser: (eg: Firefox, Chrome, incognito mode, etc, which worked, which didn't) - Client platform: (eg: Windows, MacOS, Linux) - App: (eg: webapp, admin, sentinel, api, couch2pg, medic-conf, etc) - Version: (eg: 2.15.0, 3.0.0, etc) **Additional context** This [commit](https://github.com/medic/cht-core/commit/69c59f3a02e3887bda7a5d78391e82e7244a6083) adds the new "username" parameter to the login GET request, without adding it to the ignored URL parameters in the service worker (https://github.com/medic/cht-core/blob/master/grunt/service-worker.js#L42)
1.0
Login page is not cached correctly - **Describe the bug** If a user logs off while being disconnected from the internet, they won't be able to access the login page. **To Reproduce** Steps to reproduce the behavior: 1. Login as an offline user. 2. Disconnect from the internet. 3. Logout. 4. You're seeing your browser's "No internet" page instead of the cached login page. **Expected behavior** Login page should be cached: https://github.com/medic/cht-core/pull/5557 **Environment** - Instance: (eg: alpha.dev.medicmobile.org, etc) - Browser: (eg: Firefox, Chrome, incognito mode, etc, which worked, which didn't) - Client platform: (eg: Windows, MacOS, Linux) - App: (eg: webapp, admin, sentinel, api, couch2pg, medic-conf, etc) - Version: (eg: 2.15.0, 3.0.0, etc) **Additional context** This [commit](https://github.com/medic/cht-core/commit/69c59f3a02e3887bda7a5d78391e82e7244a6083) adds the new "username" parameter to the login GET request, without adding it to the ignored URL parameters in the service worker (https://github.com/medic/cht-core/blob/master/grunt/service-worker.js#L42)
non_test
login page is not cached correctly describe the bug if a user logs off while being disconnected from the internet they won t be able to access the login page to reproduce steps to reproduce the behavior login as an offline user disconnect from the internet logout you re seeing your browser s no internet page instead of the cached login page expected behavior login page should be cached environment instance eg alpha dev medicmobile org etc browser eg firefox chrome incognito mode etc which worked which didn t client platform eg windows macos linux app eg webapp admin sentinel api medic conf etc version eg etc additional context this adds the new username parameter to the login get request without adding it to the ignored url parameters in the service worker
0
121,886
12,136,547,691
IssuesEvent
2020-04-23 14:31:58
openssl/openssl
https://api.github.com/repos/openssl/openssl
closed
CMP is not mentioned in CHANGES
triaged: documentation
CMP is not mentioned in the CHANGES file. This this is a significant new feature of 3.0 we should have something about it.
1.0
CMP is not mentioned in CHANGES - CMP is not mentioned in the CHANGES file. This this is a significant new feature of 3.0 we should have something about it.
non_test
cmp is not mentioned in changes cmp is not mentioned in the changes file this this is a significant new feature of we should have something about it
0
263,272
23,045,721,297
IssuesEvent
2022-07-23 21:45:39
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Frequent GPU crashes with Radeon 5700XT
bug topic:rendering needs testing crash
### Godot version 4.0 Alpha ### System information Windows 10 20H2, Vulkan, AMD Radeon 5700XT (21.12.1), Ryzen 3900X, 32GB RAM ### Issue description When running the editor I will get frequent GPU and system crashes. ### Steps to reproduce Open the editor for a few minutes. ### Minimal reproduction project _No response_
1.0
Frequent GPU crashes with Radeon 5700XT - ### Godot version 4.0 Alpha ### System information Windows 10 20H2, Vulkan, AMD Radeon 5700XT (21.12.1), Ryzen 3900X, 32GB RAM ### Issue description When running the editor I will get frequent GPU and system crashes. ### Steps to reproduce Open the editor for a few minutes. ### Minimal reproduction project _No response_
test
frequent gpu crashes with radeon godot version alpha system information windows vulkan amd radeon ryzen ram issue description when running the editor i will get frequent gpu and system crashes steps to reproduce open the editor for a few minutes minimal reproduction project no response
1
436,982
30,580,607,573
IssuesEvent
2023-07-21 09:23:30
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
Doc should mentioned that support vector will be empty with precomputed kernel
Documentation module:svm
#### Describe the bug After training an SVM with callable kernel function, the SVM's support vectors attribute is an empty array. #### Steps/Code to Reproduce Minimal example based on [this tutorial](https://scikit-learn.org/stable/auto_examples/svm/plot_custom_kernel.html) ```python import numpy as np from sklearn import svm, datasets # import some data to play with iris = datasets.load_iris() X = iris.data[:, :2] Y = iris.target def my_kernel(X, Y): """ We create a custom kernel: (2 0) k(X, Y) = X ( ) Y.T (0 1) """ M = np.array([[2, 0], [0, 1.0]]) return np.dot(np.dot(X, M), Y.T) # we create an instance of SVM and fit out data. clf = svm.SVC(kernel=my_kernel) clf.fit(X, Y) print(clf.support_vectors_) # array([], shape=(0, 0), dtype=float64) print(clf.n_support_) # array([ 7, 40, 34], dtype=int32) print(clf.support_) # array of length 81 ``` #### Expected Results Expected `clf.support_vectors_` to be an array of shape [num_support_vectors, num_features] #### Actual Results `clf.support_vectors_` is an array of shape [0, 0] #### Versions setuptools: 54.0.0 sklearn: 0.23.2 numpy: 1.19.2 scipy: 1.6.1 Cython: None pandas: 1.2.2 matplotlib: 3.3.4 joblib: 1.0.1 threadpoolctl: 2.1.0 Built with OpenMP: True
1.0
Doc should mentioned that support vector will be empty with precomputed kernel - #### Describe the bug After training an SVM with callable kernel function, the SVM's support vectors attribute is an empty array. #### Steps/Code to Reproduce Minimal example based on [this tutorial](https://scikit-learn.org/stable/auto_examples/svm/plot_custom_kernel.html) ```python import numpy as np from sklearn import svm, datasets # import some data to play with iris = datasets.load_iris() X = iris.data[:, :2] Y = iris.target def my_kernel(X, Y): """ We create a custom kernel: (2 0) k(X, Y) = X ( ) Y.T (0 1) """ M = np.array([[2, 0], [0, 1.0]]) return np.dot(np.dot(X, M), Y.T) # we create an instance of SVM and fit out data. clf = svm.SVC(kernel=my_kernel) clf.fit(X, Y) print(clf.support_vectors_) # array([], shape=(0, 0), dtype=float64) print(clf.n_support_) # array([ 7, 40, 34], dtype=int32) print(clf.support_) # array of length 81 ``` #### Expected Results Expected `clf.support_vectors_` to be an array of shape [num_support_vectors, num_features] #### Actual Results `clf.support_vectors_` is an array of shape [0, 0] #### Versions setuptools: 54.0.0 sklearn: 0.23.2 numpy: 1.19.2 scipy: 1.6.1 Cython: None pandas: 1.2.2 matplotlib: 3.3.4 joblib: 1.0.1 threadpoolctl: 2.1.0 Built with OpenMP: True
non_test
doc should mentioned that support vector will be empty with precomputed kernel describe the bug after training an svm with callable kernel function the svm s support vectors attribute is an empty array steps code to reproduce minimal example based on python import numpy as np from sklearn import svm datasets import some data to play with iris datasets load iris x iris data y iris target def my kernel x y we create a custom kernel k x y x y t m np array return np dot np dot x m y t we create an instance of svm and fit out data clf svm svc kernel my kernel clf fit x y print clf support vectors array shape dtype print clf n support array dtype print clf support array of length expected results expected clf support vectors to be an array of shape actual results clf support vectors is an array of shape versions setuptools sklearn numpy scipy cython none pandas matplotlib joblib threadpoolctl built with openmp true
0
17,893
3,013,568,576
IssuesEvent
2015-07-29 09:46:42
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Invalid XML returned by YEngine.getStateForCase(YIdentifier caseID)
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Create a YEngine object 2. Load a specification 3. Start a case 4. Call YEngine.getStateForCase(caseID) with the case's ID What is the expected output? What do you see instead? Expected (only concerns this one element): <nextElementRef id="foobar" documentation=""/> Received: <nextElementRef id="foobar" documentation=""> Please use labels and text to provide additional information. The slash '/' closing an empty XML tag is missing. It needs to be added in YEngine.class:999. The problem exists in trunk/release2/src/org/yawlfoundation/yawl/engine/YEngine.java revision 1798. Don't know about other revisions. ``` Original issue reported on code.google.com by `florian....@gmail.com` on 5 Dec 2011 at 3:13
1.0
Invalid XML returned by YEngine.getStateForCase(YIdentifier caseID) - ``` What steps will reproduce the problem? 1. Create a YEngine object 2. Load a specification 3. Start a case 4. Call YEngine.getStateForCase(caseID) with the case's ID What is the expected output? What do you see instead? Expected (only concerns this one element): <nextElementRef id="foobar" documentation=""/> Received: <nextElementRef id="foobar" documentation=""> Please use labels and text to provide additional information. The slash '/' closing an empty XML tag is missing. It needs to be added in YEngine.class:999. The problem exists in trunk/release2/src/org/yawlfoundation/yawl/engine/YEngine.java revision 1798. Don't know about other revisions. ``` Original issue reported on code.google.com by `florian....@gmail.com` on 5 Dec 2011 at 3:13
non_test
invalid xml returned by yengine getstateforcase yidentifier caseid what steps will reproduce the problem create a yengine object load a specification start a case call yengine getstateforcase caseid with the case s id what is the expected output what do you see instead expected only concerns this one element nextelementref id foobar documentation received please use labels and text to provide additional information the slash closing an empty xml tag is missing it needs to be added in yengine class the problem exists in trunk src org yawlfoundation yawl engine yengine java revision don t know about other revisions original issue reported on code google com by florian gmail com on dec at
0
133,893
18,363,445,273
IssuesEvent
2021-10-09 16:33:22
tuanducdesign/serviceapp
https://api.github.com/repos/tuanducdesign/serviceapp
opened
CVE-2018-19797 (Medium) detected in node-sass-4.14.0.tgz, opennmsopennms-source-26.0.0-1
security vulnerability
## CVE-2018-19797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.0.tgz</b>, <b>opennmsopennms-source-26.0.0-1</b></p></summary> <p> <details><summary><b>node-sass-4.14.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.0.tgz</a></p> <p>Path to dependency file: serviceapp/package.json</p> <p>Path to vulnerable library: /node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/serviceapp/commit/9c49706d08f6181a261d95ef013b335d21707fb3">9c49706d08f6181a261d95ef013b335d21707fb3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p> <p>Release Date: 2018-12-03</p> <p>Fix Resolution: libsass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19797 (Medium) detected in node-sass-4.14.0.tgz, opennmsopennms-source-26.0.0-1 - ## CVE-2018-19797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.0.tgz</b>, <b>opennmsopennms-source-26.0.0-1</b></p></summary> <p> <details><summary><b>node-sass-4.14.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.0.tgz</a></p> <p>Path to dependency file: serviceapp/package.json</p> <p>Path to vulnerable library: /node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/serviceapp/commit/9c49706d08f6181a261d95ef013b335d21707fb3">9c49706d08f6181a261d95ef013b335d21707fb3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p> <p>Release Date: 2018-12-03</p> <p>Fix Resolution: libsass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in node sass tgz opennmsopennms source cve medium severity vulnerability vulnerable libraries node sass tgz opennmsopennms source node sass tgz wrapper around libsass library home page a href path to dependency file serviceapp package json path to vulnerable library node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass a null pointer dereference in the function sass selector list populate extends in sharedptr hpp used by ast cpp and ast selectors cpp may cause a denial of service application crash via a crafted sass input file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
90,250
8,231,146,457
IssuesEvent
2018-09-07 15:02:25
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
util/ctxgroup: (unknown) failed under stress
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/517663fa1489a2f36a593dc9722384a50bfcdbb5 Parameters: ``` TAGS= GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=891198&tab=buildLog ``` find ./pkg -name node_modules -prune -o -type f -name '*.pb.go' -exec rm {} + github.com/cockroachdb/cockroach/vendor/github.com/golang/glog github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/go/gcexportdata yes set -e; for dir in ./pkg/acceptance/cluster/ ./pkg/build/ ./pkg/ccl/backupccl/ ./pkg/ccl/baseccl/ ./pkg/ccl/storageccl/engineccl/enginepbccl/ ./pkg/ccl/utilccl/licenseccl/ ./pkg/config/ ./pkg/gossip/ ./pkg/internal/client/ ./pkg/roachpb/ ./pkg/rpc/ ./pkg/server/diagnosticspb/ ./pkg/server/serverpb/ ./pkg/server/status/ ./pkg/settings/cluster/ ./pkg/sql/distsqlrun/ ./pkg/sql/jobs/ ./pkg/sql/pgwire/pgerror/ ./pkg/sql/sqlbase/ ./pkg/sql/stats/ ./pkg/storage/ ./pkg/storage/engine/enginepb/ ./pkg/storage/storagebase/ ./pkg/ts/tspb/ ./pkg/util/ ./pkg/util/hlc/ ./pkg/util/log/ ./pkg/util/protoutil/ ./pkg/util/tracing/; do \ build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --plugin=bin/protoc-gen-gogoroach --gogoroach_out=Mgoogle/api/annotations.proto=github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis/google/api,Mgoogle/protobuf/timestamp.proto=github.com/gogo/protobuf/types,,plugins=grpc,import_prefix=github.com/cockroachdb/cockroach/pkg/:./pkg $dir/*.proto; \ done touch bin/.cpp_ccl_protobuf_sources checking for sys/types.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/lint yes sed -i -E '/gogoproto/d' /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.cc github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/plugin github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/httprule github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata touch bin/.cpp_protobuf_sources github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck/internal/errcheck checking for sys/stat.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/generator github.com/cockroachdb/cockroach/vendor/github.com/golang/lint/golint github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cover github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata/go-bindata github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck yes github.com/cockroachdb/cockroach/vendor/github.com/mattn/goveralls checking for stdlib.h... github.com/cockroachdb/cockroach/vendor/github.com/client9/misspell/cmd/misspell Scanning dependencies of target roach [ 2%] Building CXX object CMakeFiles/roach.dir/batch.cc.o yes [ 5%] Building CXX object CMakeFiles/roach.dir/comparator.cc.o [ 8%] Building CXX object CMakeFiles/roach.dir/encoding.cc.o [ 11%] Building CXX object CMakeFiles/roach.dir/chunked_buffer.cc.o [ 14%] Building CXX object CMakeFiles/roach.dir/db.cc.o checking for string.h... yes checking for memory.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/suffixtree yes github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/descriptor github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax checking for strings.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax/golang yes github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/job github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/output checking for inttypes.h... github.com/cockroachdb/cockroach/vendor/github.com/wadey/gocovmerge github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl github.com/cockroachdb/cockroach/vendor/github.com/golang/dep yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/feedback github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/generator checking for stdint.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/internal/stats github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/storage/benchfmt github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/gengateway github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/imports yes checking for unistd.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/benchstat github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/base yes github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/cmd/benchstat github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glide github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glock github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/godep checking whether byte ordering is bigendian... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govend github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govendor github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/gvt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/vndr github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goyacc go/internal/gccgoimporter go/internal/gcimporter go/internal/srcimporter github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goimports github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/cmd/dep no checking size of void *... go/importer github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/stringer 8 checking size of int... 4 checking size of long... 8 checking size of long long... 8 checking size of intmax_t... 8 checking build system type... [ 17%] Building CXX object CMakeFiles/roach.dir/engine.cc.o x86_64-pc-linux-gnu checking host system type... x86_64-pc-linux-gnu checking whether pause instruction is compilable... yes checking for ar... ar checking malloc.h usability... yes checking malloc.h presence... yes checking for malloc.h... yes checking whether malloc_usable_size definition can use const argument... no [ 20%] Building CXX object CMakeFiles/roach.dir/merge.cc.o checking for library containing log... [ 23%] Building CXX object CMakeFiles/roach.dir/mvcc.cc.o -lm checking whether __attribute__ syntax is compilable... yes checking whether compiler supports -fvisibility=hidden... [ 26%] Building CXX object CMakeFiles/roach.dir/options.cc.o yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether tls_model attribute is compilable... touch bin/.bootstrap [ 29%] Building CXX object CMakeFiles/roach.dir/snapshot.cc.o yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether alloc_size attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(gnu_printf, ...) attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(printf, ...) attribute is compilable... yes checking for a BSD-compatible install... /usr/bin/install -c checking for ranlib... ranlib checking for ld... /usr/bin/ld checking for autoconf... /usr/bin/autoconf checking for memalign... yes checking for valloc... yes checking whether compiler supports -O3... yes checking whether compiler supports -funroll-loops... yes checking unwind.h usability... yes checking unwind.h presence... yes checking for unwind.h... yes checking for _Unwind_Backtrace in -lgcc... yes checking configured backtracing method... libgcc checking for sbrk... yes checking whether utrace(2) is compilable... no checking whether valgrind is compilable... no checking whether a program using __builtin_unreachable is compilable... yes checking whether a program using __builtin_ffsl is compilable... yes checking LG_PAGE... 12 Missing VERSION file, and unable to generate it; creating bogus VERSION [ 32%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/data.pb.cc.o checking pthread.h usability... yes checking pthread.h presence... yes checking for pthread.h... yes checking for pthread_create in -lpthread... yes checking whether pthread_atfork(3) is compilable... yes checking for library containing clock_gettime... none required checking whether clock_gettime(CLOCK_MONOTONIC_COARSE, ...) is compilable... yes checking whether clock_gettime(CLOCK_MONOTONIC, ...) is compilable... yes checking whether mach_absolute_time() is compilable... no checking whether compiler supports -Werror... yes checking whether syscall(2) is compilable... yes checking for secure_getenv... yes checking for issetugid... no checking for _malloc_thread_cleanup... [ 35%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/internal.pb.cc.o no checking for _pthread_mutex_init_calloc_cb... no checking for TLS... yes checking whether C11 atomics is compilable... [ 38%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/metadata.pb.cc.o yes checking whether atomic(9) is compilable... no checking whether Darwin OSAtomic*() is compilable... no checking whether madvise(2) is compilable... yes checking whether madvise(..., MADV_FREE) is compilable... [ 41%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc.pb.cc.o no checking whether madvise(..., MADV_DONTNEED) is compilable... [ 44%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc3.pb.cc.o yes checking whether madvise(..., MADV_[NO]HUGEPAGE) is compilable... [ 47%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/file_registry.pb.cc.o yes checking whether to force 32-bit __sync_{add,sub}_and_fetch()... no checking whether to force 64-bit __sync_{add,sub}_and_fetch()... no checking for __builtin_clz... yes checking whether Darwin os_unfair_lock_*() is compilable... no checking whether Darwin OSSpin*() is compilable... no checking whether glibc malloc hook is compilable... yes checking whether glibc memalign hook is compilable... yes checking whether pthreads adaptive mutexes is compilable... yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes configure: creating ./config.status config.status: creating Makefile config.status: creating jemalloc.pc config.status: creating doc/html.xsl config.status: creating doc/manpages.xsl config.status: creating doc/jemalloc.xml config.status: creating include/jemalloc/jemalloc_macros.h config.status: creating include/jemalloc/jemalloc_protos.h [ 50%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/rocksdb.pb.cc.o config.status: creating include/jemalloc/jemalloc_typedefs.h config.status: creating include/jemalloc/internal/jemalloc_internal.h config.status: creating test/test.sh config.status: creating test/include/test/jemalloc_test.h config.status: creating config.stamp config.status: creating bin/jemalloc-config config.status: creating bin/jemalloc.sh config.status: creating bin/jeprof config.status: creating include/jemalloc/jemalloc_defs.h config.status: creating include/jemalloc/internal/jemalloc_internal_defs.h config.status: creating test/include/test/jemalloc_test_defs.h config.status: executing include/jemalloc/internal/private_namespace.h commands config.status: executing include/jemalloc/internal/private_unnamespace.h commands config.status: executing include/jemalloc/internal/public_symbols.txt commands config.status: executing include/jemalloc/internal/public_namespace.h commands [ 52%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/legacy_timestamp.pb.cc.o config.status: executing include/jemalloc/internal/public_unnamespace.h commands [ 55%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/timestamp.pb.cc.o config.status: executing include/jemalloc/internal/size_classes.h commands [ 58%] Building CXX object CMakeFiles/roach.dir/protos/util/unresolved_addr.pb.cc.o config.status: executing include/jemalloc/jemalloc_protos_jet.h commands config.status: executing include/jemalloc/jemalloc_rename.h commands config.status: executing include/jemalloc/jemalloc_mangle.h commands set -euo pipefail; \ ret=$(cd ./pkg/sql/parser/gen && goyacc -p sql -o sql.go.tmp sql.y); \ if expr "$ret" : ".*conflicts" >/dev/null; then \ echo "$ret"; exit 1; \ fi config.status: executing include/jemalloc/jemalloc_mangle_jet.h commands config.status: executing include/jemalloc/jemalloc.h commands =============================================================================== jemalloc version : 0.0.0-0-g0000000000000000000000000000000000000000 library revision : 2 CONFIG : --enable-prof CC : cc CONFIGURE_CFLAGS : -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops SPECIFIED_CFLAGS : EXTRA_CFLAGS : CPPFLAGS : -D_GNU_SOURCE -D_REENTRANT LDFLAGS : EXTRA_LDFLAGS : LIBS : -lm -lgcc -lm -lpthread RPATH_EXTRA : XSLTPROC : false XSLROOT : PREFIX : /usr/local BINDIR : /usr/local/bin DATADIR : /usr/local/share INCLUDEDIR : /usr/local/include LIBDIR : /usr/local/lib MANDIR : /usr/local/share/man srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ abs_srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ objroot : abs_objroot : /go/native/x86_64-pc-linux-gnu/jemalloc/ JEMALLOC_PREFIX : JEMALLOC_PRIVATE_NAMESPACE : je_ install_suffix : malloc_conf : autogen : 0 cc-silence : 1 debug : 0 code-coverage : 0 stats : 1 prof : 1 prof-libunwind : 0 prof-libgcc : 1 prof-gcc : 0 tcache : 1 thp : 1 fill : 1 utrace : 0 valgrind : 0 xmalloc : 0 munmap : 0 lazy_lock : 0 tls : 1 cache-oblivious : 1 =============================================================================== find ./pkg -name node_modules -prune -o -type f -name '*.pb.gw.go' -exec rm {} + build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/server/serverpb/admin.proto ./pkg/server/serverpb/status.proto ./pkg/server/serverpb/authentication.proto build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/ts/tspb/timeseries.proto sed -i '/import _/d' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go gofmt -s -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go goimports -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go [ 61%] Linking CXX static library libroach.a sed -i -E 's!import (fmt|math) "github.com/cockroachdb/cockroach/pkg/(fmt|math)"! !g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go [ 82%] Built target roach Scanning dependencies of target roachccl [ 88%] Building CXX object CMakeFiles/roachccl.dir/ccl/db.cc.o [ 88%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc.o [ 91%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/baseccl/encryption_options.pb.cc.o [ 94%] Building CXX object CMakeFiles/roachccl.dir/ccl/key_manager.cc.o touch bin/.gw_protobuf_sources sed -i -E 's!cockroachdb/cockroach/pkg/(etcd)!coreos/\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!github.com/cockroachdb/cockroach/pkg/(bytes|encoding/binary|errors|fmt|io|math|github\.com|(google\.)?golang\.org)!\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go gofmt -s -w ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go touch bin/.go_protobuf_sources cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidde(echo "// Code generated by goyacc. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ cat pkg/sql/parser/gen/sql.go.tmp | \ sed -E 's/^const ([A-Z][_A-Z0-9]*) =.*$/const \1 = lex.\1/g') > pkg/sql/parser/sql.go (echo "// Code generated by make. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ echo; \ echo "package lex"; \ echo; \ grep '^const [A-Z][_A-Z0-9]* ' pkg/sql/parser/gen/sql.go.tmp) > pkg/sql/lex/tokens.go go run pkg/cmd/docgen/{main,funcs}.go functions docs/generated/sql --quiet [ 97%] Linking CXX static library libroachccl.a [100%] Built target roachccl n -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/sar: `u' modifier ignored since `D' is the default (see `U') ar: `u' modifier ignored since `D' is the default (see `U') rc/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c ar crus lib/libjemalloc.a src/jemalloc.o src/arena.o src/atomic.o src/base.o src/bitmap.o src/chunk.o src/chunk_dss.o src/chunk_mmap.o src/ckh.o src/ctl.o src/extent.o src/hash.o src/huge.o src/mb.o src/mutex.o src/nstime.o src/pages.o src/prng.o src/prof.o src/quarantine.o src/rtree.o src/stats.o src/spin.o src/tcache.o src/ticker.o src/tsd.o src/util.o src/witness.o ar crus lib/libjemalloc_pic.a src/jemalloc.pic.o src/arena.pic.o src/atomic.pic.o src/base.pic.o src/bitmap.pic.o src/chunk.pic.o src/chunk_dss.pic.o src/chunk_mmap.pic.o src/ckh.pic.o src/ctl.pic.o src/extent.pic.o src/hash.pic.o src/huge.pic.o src/mb.pic.o src/mutex.pic.o src/nstime.pic.o src/pages.pic.o src/prng.pic.o src/prof.pic.o src/quarantine.pic.o src/rtree.pic.o src/stats.pic.o src/spin.pic.o src/tcache.pic.o src/ticker.pic.o src/tsd.pic.o src/util.pic.o src/witness.pic.o [ 0%] Built target build_version touch bin/.docgen_functions Scanning dependencies of target rocksdb [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/clock_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/lru_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/sharded_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/c.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/builder.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/column_family.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compacted_db_impl.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_iterator.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_job.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker_universal.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/convenience.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_filesnapshot.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_write.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_compaction_flush.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_files.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_open.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_debug.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_readonly.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_info_dumper.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_iter.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/dbformat.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/event_helpers.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/external_sst_file_ingestion_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/file_indexer.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_scheduler.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/forward_iterator.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/internal_stats.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_reader.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_writer.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/malloc_stats.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/managed_iterator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable_list.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/merge_helper.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/range_del_aggregator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/repair.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_cache.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_properties_collector.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/transaction_log_impl.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_builder.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_edit.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_set.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/wal_manager.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_batch.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_thread.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/env/env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/env_encryption.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/mock_env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/alloc_tracker.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_cuckoo_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_linklist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_skiplist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/skiplistrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/vectorrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/write_buffer_manager.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/histogram.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/instrumented_mutex.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/iostats_context.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_context.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_level.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/statistics.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_updater.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util_debug.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/options/cf_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/db_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_helper.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_parser.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/adaptive_table_factory.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_filter_block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_factory.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_reader.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_prefix_index.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/bloom_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_builder.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_factory.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_reader.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/format.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/full_filter_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/get_context.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/index_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/merging_iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/meta_blocks.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/partitioned_filter_block.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/persistent_cache_helper.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_factory.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_index.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_key_coding.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_reader.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/sst_file_writer.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/table_properties.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/two_level_iterator.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/tools/db_bench_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/dump/db_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/ldb_cmd.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/sst_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/arena.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/auto_roll_logger.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/coding.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/comparator.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/concurrent_arena.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/crc32c.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/delete_scheduler.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/dynamic_bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/event_logger.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_reader_writer.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_util.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/filename.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/hash.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/log_buffer.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/random.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/rate_limiter.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sst_file_manager_impl.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/status.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sync_point.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/testutil.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/thread_local.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/threadpool_imp.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/transaction_test_util.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/backupable/backupable_db.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_dump_tool.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_file.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_reader.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_writer.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_format.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/ttl_extractor.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/checkpoint/checkpoint_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_decoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_encoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/column_aware_encoding_util.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/date_tiered/date_tiered_db_impl.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/debug.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/document_db.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/json_document.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/env_timed.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/geodb/geodb_impl.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/memory/memory_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/merge_operators/uint64add.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/options/options_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_file.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_metadata.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/volatile_tier_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/redis/redis_lists.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/simulator_cache/sim_cache.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/spatialdb/spatial_db.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction_db_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction_db.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/snapshot_checker.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_base.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_lock_mgr.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_util.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/write_prepared_txn.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/ttl/db_ttl_impl.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index_internal.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/port/port_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/env_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/io_posix.cc.o [ 89%] Linking CXX static library librocksdb.a [100%] Built target rocksdb go test -v -tags ' make x86_64_pc_linux_gnu' -ldflags ' -X github.com/cockroachdb/cockroach/pkg/build.typ=development -X "github.com/cockroachdb/cockroach/pkg/build.tag=v2.0.5-34-g517663f-dirty" -X "github.com/cockroachdb/cockroach/pkg/build.utcTime=2018/09/07 08:29:04" -X "github.com/cockroachdb/cockroach/pkg/build.rev=517663fa1489a2f36a593dc9722384a50bfcdbb5" -X "github.com/cockroachdb/cockroach/pkg/build.cgoTargetTriple=x86_64-pc-linux-gnu" ' -i github.com/cockroachdb/cockroach/pkg/util/ctxgroup can't load package: package github.com/cockroachdb/cockroach/pkg/util/ctxgroup: cannot find package "github.com/cockroachdb/cockroach/pkg/util/ctxgroup" in any of: /usr/local/go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup (from $GOROOT) /go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup (from $GOPATH) Makefile:756: recipe for target 'gotestdashi' failed make: *** [gotestdashi] Error 1 ```
1.0
util/ctxgroup: (unknown) failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/517663fa1489a2f36a593dc9722384a50bfcdbb5 Parameters: ``` TAGS= GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=891198&tab=buildLog ``` find ./pkg -name node_modules -prune -o -type f -name '*.pb.go' -exec rm {} + github.com/cockroachdb/cockroach/vendor/github.com/golang/glog github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/go/gcexportdata yes set -e; for dir in ./pkg/acceptance/cluster/ ./pkg/build/ ./pkg/ccl/backupccl/ ./pkg/ccl/baseccl/ ./pkg/ccl/storageccl/engineccl/enginepbccl/ ./pkg/ccl/utilccl/licenseccl/ ./pkg/config/ ./pkg/gossip/ ./pkg/internal/client/ ./pkg/roachpb/ ./pkg/rpc/ ./pkg/server/diagnosticspb/ ./pkg/server/serverpb/ ./pkg/server/status/ ./pkg/settings/cluster/ ./pkg/sql/distsqlrun/ ./pkg/sql/jobs/ ./pkg/sql/pgwire/pgerror/ ./pkg/sql/sqlbase/ ./pkg/sql/stats/ ./pkg/storage/ ./pkg/storage/engine/enginepb/ ./pkg/storage/storagebase/ ./pkg/ts/tspb/ ./pkg/util/ ./pkg/util/hlc/ ./pkg/util/log/ ./pkg/util/protoutil/ ./pkg/util/tracing/; do \ build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --plugin=bin/protoc-gen-gogoroach --gogoroach_out=Mgoogle/api/annotations.proto=github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis/google/api,Mgoogle/protobuf/timestamp.proto=github.com/gogo/protobuf/types,,plugins=grpc,import_prefix=github.com/cockroachdb/cockroach/pkg/:./pkg $dir/*.proto; \ done touch bin/.cpp_ccl_protobuf_sources checking for sys/types.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/lint yes sed -i -E '/gogoproto/d' /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.cc github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/plugin github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/httprule github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata touch bin/.cpp_protobuf_sources github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck/internal/errcheck checking for sys/stat.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/generator github.com/cockroachdb/cockroach/vendor/github.com/golang/lint/golint github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cover github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata/go-bindata github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck yes github.com/cockroachdb/cockroach/vendor/github.com/mattn/goveralls checking for stdlib.h... github.com/cockroachdb/cockroach/vendor/github.com/client9/misspell/cmd/misspell Scanning dependencies of target roach [ 2%] Building CXX object CMakeFiles/roach.dir/batch.cc.o yes [ 5%] Building CXX object CMakeFiles/roach.dir/comparator.cc.o [ 8%] Building CXX object CMakeFiles/roach.dir/encoding.cc.o [ 11%] Building CXX object CMakeFiles/roach.dir/chunked_buffer.cc.o [ 14%] Building CXX object CMakeFiles/roach.dir/db.cc.o checking for string.h... yes checking for memory.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/suffixtree yes github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/descriptor github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax checking for strings.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax/golang yes github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/job github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/output checking for inttypes.h... github.com/cockroachdb/cockroach/vendor/github.com/wadey/gocovmerge github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl github.com/cockroachdb/cockroach/vendor/github.com/golang/dep yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/feedback github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/generator checking for stdint.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/internal/stats github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/storage/benchfmt github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/gengateway github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/imports yes checking for unistd.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/benchstat github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/base yes github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/cmd/benchstat github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glide github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glock github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/godep checking whether byte ordering is bigendian... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govend github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govendor github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/gvt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/vndr github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goyacc go/internal/gccgoimporter go/internal/gcimporter go/internal/srcimporter github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goimports github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/cmd/dep no checking size of void *... go/importer github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/stringer 8 checking size of int... 4 checking size of long... 8 checking size of long long... 8 checking size of intmax_t... 8 checking build system type... [ 17%] Building CXX object CMakeFiles/roach.dir/engine.cc.o x86_64-pc-linux-gnu checking host system type... x86_64-pc-linux-gnu checking whether pause instruction is compilable... yes checking for ar... ar checking malloc.h usability... yes checking malloc.h presence... yes checking for malloc.h... yes checking whether malloc_usable_size definition can use const argument... no [ 20%] Building CXX object CMakeFiles/roach.dir/merge.cc.o checking for library containing log... [ 23%] Building CXX object CMakeFiles/roach.dir/mvcc.cc.o -lm checking whether __attribute__ syntax is compilable... yes checking whether compiler supports -fvisibility=hidden... [ 26%] Building CXX object CMakeFiles/roach.dir/options.cc.o yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether tls_model attribute is compilable... touch bin/.bootstrap [ 29%] Building CXX object CMakeFiles/roach.dir/snapshot.cc.o yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether alloc_size attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(gnu_printf, ...) attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(printf, ...) attribute is compilable... yes checking for a BSD-compatible install... /usr/bin/install -c checking for ranlib... ranlib checking for ld... /usr/bin/ld checking for autoconf... /usr/bin/autoconf checking for memalign... yes checking for valloc... yes checking whether compiler supports -O3... yes checking whether compiler supports -funroll-loops... yes checking unwind.h usability... yes checking unwind.h presence... yes checking for unwind.h... yes checking for _Unwind_Backtrace in -lgcc... yes checking configured backtracing method... libgcc checking for sbrk... yes checking whether utrace(2) is compilable... no checking whether valgrind is compilable... no checking whether a program using __builtin_unreachable is compilable... yes checking whether a program using __builtin_ffsl is compilable... yes checking LG_PAGE... 12 Missing VERSION file, and unable to generate it; creating bogus VERSION [ 32%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/data.pb.cc.o checking pthread.h usability... yes checking pthread.h presence... yes checking for pthread.h... yes checking for pthread_create in -lpthread... yes checking whether pthread_atfork(3) is compilable... yes checking for library containing clock_gettime... none required checking whether clock_gettime(CLOCK_MONOTONIC_COARSE, ...) is compilable... yes checking whether clock_gettime(CLOCK_MONOTONIC, ...) is compilable... yes checking whether mach_absolute_time() is compilable... no checking whether compiler supports -Werror... yes checking whether syscall(2) is compilable... yes checking for secure_getenv... yes checking for issetugid... no checking for _malloc_thread_cleanup... [ 35%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/internal.pb.cc.o no checking for _pthread_mutex_init_calloc_cb... no checking for TLS... yes checking whether C11 atomics is compilable... [ 38%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/metadata.pb.cc.o yes checking whether atomic(9) is compilable... no checking whether Darwin OSAtomic*() is compilable... no checking whether madvise(2) is compilable... yes checking whether madvise(..., MADV_FREE) is compilable... [ 41%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc.pb.cc.o no checking whether madvise(..., MADV_DONTNEED) is compilable... [ 44%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc3.pb.cc.o yes checking whether madvise(..., MADV_[NO]HUGEPAGE) is compilable... [ 47%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/file_registry.pb.cc.o yes checking whether to force 32-bit __sync_{add,sub}_and_fetch()... no checking whether to force 64-bit __sync_{add,sub}_and_fetch()... no checking for __builtin_clz... yes checking whether Darwin os_unfair_lock_*() is compilable... no checking whether Darwin OSSpin*() is compilable... no checking whether glibc malloc hook is compilable... yes checking whether glibc memalign hook is compilable... yes checking whether pthreads adaptive mutexes is compilable... yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes configure: creating ./config.status config.status: creating Makefile config.status: creating jemalloc.pc config.status: creating doc/html.xsl config.status: creating doc/manpages.xsl config.status: creating doc/jemalloc.xml config.status: creating include/jemalloc/jemalloc_macros.h config.status: creating include/jemalloc/jemalloc_protos.h [ 50%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/rocksdb.pb.cc.o config.status: creating include/jemalloc/jemalloc_typedefs.h config.status: creating include/jemalloc/internal/jemalloc_internal.h config.status: creating test/test.sh config.status: creating test/include/test/jemalloc_test.h config.status: creating config.stamp config.status: creating bin/jemalloc-config config.status: creating bin/jemalloc.sh config.status: creating bin/jeprof config.status: creating include/jemalloc/jemalloc_defs.h config.status: creating include/jemalloc/internal/jemalloc_internal_defs.h config.status: creating test/include/test/jemalloc_test_defs.h config.status: executing include/jemalloc/internal/private_namespace.h commands config.status: executing include/jemalloc/internal/private_unnamespace.h commands config.status: executing include/jemalloc/internal/public_symbols.txt commands config.status: executing include/jemalloc/internal/public_namespace.h commands [ 52%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/legacy_timestamp.pb.cc.o config.status: executing include/jemalloc/internal/public_unnamespace.h commands [ 55%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/timestamp.pb.cc.o config.status: executing include/jemalloc/internal/size_classes.h commands [ 58%] Building CXX object CMakeFiles/roach.dir/protos/util/unresolved_addr.pb.cc.o config.status: executing include/jemalloc/jemalloc_protos_jet.h commands config.status: executing include/jemalloc/jemalloc_rename.h commands config.status: executing include/jemalloc/jemalloc_mangle.h commands set -euo pipefail; \ ret=$(cd ./pkg/sql/parser/gen && goyacc -p sql -o sql.go.tmp sql.y); \ if expr "$ret" : ".*conflicts" >/dev/null; then \ echo "$ret"; exit 1; \ fi config.status: executing include/jemalloc/jemalloc_mangle_jet.h commands config.status: executing include/jemalloc/jemalloc.h commands =============================================================================== jemalloc version : 0.0.0-0-g0000000000000000000000000000000000000000 library revision : 2 CONFIG : --enable-prof CC : cc CONFIGURE_CFLAGS : -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops SPECIFIED_CFLAGS : EXTRA_CFLAGS : CPPFLAGS : -D_GNU_SOURCE -D_REENTRANT LDFLAGS : EXTRA_LDFLAGS : LIBS : -lm -lgcc -lm -lpthread RPATH_EXTRA : XSLTPROC : false XSLROOT : PREFIX : /usr/local BINDIR : /usr/local/bin DATADIR : /usr/local/share INCLUDEDIR : /usr/local/include LIBDIR : /usr/local/lib MANDIR : /usr/local/share/man srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ abs_srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ objroot : abs_objroot : /go/native/x86_64-pc-linux-gnu/jemalloc/ JEMALLOC_PREFIX : JEMALLOC_PRIVATE_NAMESPACE : je_ install_suffix : malloc_conf : autogen : 0 cc-silence : 1 debug : 0 code-coverage : 0 stats : 1 prof : 1 prof-libunwind : 0 prof-libgcc : 1 prof-gcc : 0 tcache : 1 thp : 1 fill : 1 utrace : 0 valgrind : 0 xmalloc : 0 munmap : 0 lazy_lock : 0 tls : 1 cache-oblivious : 1 =============================================================================== find ./pkg -name node_modules -prune -o -type f -name '*.pb.gw.go' -exec rm {} + build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/server/serverpb/admin.proto ./pkg/server/serverpb/status.proto ./pkg/server/serverpb/authentication.proto build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/ts/tspb/timeseries.proto sed -i '/import _/d' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go gofmt -s -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go goimports -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go [ 61%] Linking CXX static library libroach.a sed -i -E 's!import (fmt|math) "github.com/cockroachdb/cockroach/pkg/(fmt|math)"! !g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go [ 82%] Built target roach Scanning dependencies of target roachccl [ 88%] Building CXX object CMakeFiles/roachccl.dir/ccl/db.cc.o [ 88%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc.o [ 91%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/baseccl/encryption_options.pb.cc.o [ 94%] Building CXX object CMakeFiles/roachccl.dir/ccl/key_manager.cc.o touch bin/.gw_protobuf_sources sed -i -E 's!cockroachdb/cockroach/pkg/(etcd)!coreos/\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!github.com/cockroachdb/cockroach/pkg/(bytes|encoding/binary|errors|fmt|io|math|github\.com|(google\.)?golang\.org)!\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go gofmt -s -w ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go touch bin/.go_protobuf_sources cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidde(echo "// Code generated by goyacc. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ cat pkg/sql/parser/gen/sql.go.tmp | \ sed -E 's/^const ([A-Z][_A-Z0-9]*) =.*$/const \1 = lex.\1/g') > pkg/sql/parser/sql.go (echo "// Code generated by make. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ echo; \ echo "package lex"; \ echo; \ grep '^const [A-Z][_A-Z0-9]* ' pkg/sql/parser/gen/sql.go.tmp) > pkg/sql/lex/tokens.go go run pkg/cmd/docgen/{main,funcs}.go functions docs/generated/sql --quiet [ 97%] Linking CXX static library libroachccl.a [100%] Built target roachccl n -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/sar: `u' modifier ignored since `D' is the default (see `U') ar: `u' modifier ignored since `D' is the default (see `U') rc/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c ar crus lib/libjemalloc.a src/jemalloc.o src/arena.o src/atomic.o src/base.o src/bitmap.o src/chunk.o src/chunk_dss.o src/chunk_mmap.o src/ckh.o src/ctl.o src/extent.o src/hash.o src/huge.o src/mb.o src/mutex.o src/nstime.o src/pages.o src/prng.o src/prof.o src/quarantine.o src/rtree.o src/stats.o src/spin.o src/tcache.o src/ticker.o src/tsd.o src/util.o src/witness.o ar crus lib/libjemalloc_pic.a src/jemalloc.pic.o src/arena.pic.o src/atomic.pic.o src/base.pic.o src/bitmap.pic.o src/chunk.pic.o src/chunk_dss.pic.o src/chunk_mmap.pic.o src/ckh.pic.o src/ctl.pic.o src/extent.pic.o src/hash.pic.o src/huge.pic.o src/mb.pic.o src/mutex.pic.o src/nstime.pic.o src/pages.pic.o src/prng.pic.o src/prof.pic.o src/quarantine.pic.o src/rtree.pic.o src/stats.pic.o src/spin.pic.o src/tcache.pic.o src/ticker.pic.o src/tsd.pic.o src/util.pic.o src/witness.pic.o [ 0%] Built target build_version touch bin/.docgen_functions Scanning dependencies of target rocksdb [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/clock_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/lru_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/sharded_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/c.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/builder.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/column_family.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compacted_db_impl.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_iterator.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_job.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker_universal.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/convenience.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_filesnapshot.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_write.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_compaction_flush.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_files.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_open.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_debug.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_readonly.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_info_dumper.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_iter.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/dbformat.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/event_helpers.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/external_sst_file_ingestion_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/file_indexer.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_scheduler.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/forward_iterator.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/internal_stats.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_reader.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_writer.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/malloc_stats.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/managed_iterator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable_list.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/merge_helper.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/range_del_aggregator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/repair.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_cache.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_properties_collector.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/transaction_log_impl.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_builder.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_edit.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_set.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/wal_manager.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_batch.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_thread.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/env/env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/env_encryption.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/mock_env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/alloc_tracker.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_cuckoo_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_linklist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_skiplist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/skiplistrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/vectorrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/write_buffer_manager.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/histogram.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/instrumented_mutex.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/iostats_context.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_context.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_level.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/statistics.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_updater.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util_debug.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/options/cf_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/db_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_helper.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_parser.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/adaptive_table_factory.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_filter_block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_factory.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_reader.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_prefix_index.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/bloom_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_builder.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_factory.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_reader.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/format.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/full_filter_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/get_context.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/index_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/merging_iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/meta_blocks.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/partitioned_filter_block.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/persistent_cache_helper.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_factory.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_index.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_key_coding.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_reader.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/sst_file_writer.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/table_properties.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/two_level_iterator.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/tools/db_bench_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/dump/db_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/ldb_cmd.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/sst_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/arena.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/auto_roll_logger.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/coding.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/comparator.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/concurrent_arena.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/crc32c.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/delete_scheduler.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/dynamic_bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/event_logger.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_reader_writer.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_util.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/filename.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/hash.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/log_buffer.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/random.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/rate_limiter.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sst_file_manager_impl.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/status.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sync_point.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/testutil.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/thread_local.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/threadpool_imp.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/transaction_test_util.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/backupable/backupable_db.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_dump_tool.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_file.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_reader.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_writer.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_format.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/ttl_extractor.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/checkpoint/checkpoint_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_decoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_encoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/column_aware_encoding_util.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/date_tiered/date_tiered_db_impl.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/debug.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/document_db.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/json_document.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/env_timed.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/geodb/geodb_impl.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/memory/memory_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/merge_operators/uint64add.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/options/options_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_file.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_metadata.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/volatile_tier_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/redis/redis_lists.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/simulator_cache/sim_cache.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/spatialdb/spatial_db.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction_db_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction_db.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/snapshot_checker.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_base.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_lock_mgr.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_util.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/write_prepared_txn.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/ttl/db_ttl_impl.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index_internal.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/port/port_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/env_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/io_posix.cc.o [ 89%] Linking CXX static library librocksdb.a [100%] Built target rocksdb go test -v -tags ' make x86_64_pc_linux_gnu' -ldflags ' -X github.com/cockroachdb/cockroach/pkg/build.typ=development -X "github.com/cockroachdb/cockroach/pkg/build.tag=v2.0.5-34-g517663f-dirty" -X "github.com/cockroachdb/cockroach/pkg/build.utcTime=2018/09/07 08:29:04" -X "github.com/cockroachdb/cockroach/pkg/build.rev=517663fa1489a2f36a593dc9722384a50bfcdbb5" -X "github.com/cockroachdb/cockroach/pkg/build.cgoTargetTriple=x86_64-pc-linux-gnu" ' -i github.com/cockroachdb/cockroach/pkg/util/ctxgroup can't load package: package github.com/cockroachdb/cockroach/pkg/util/ctxgroup: cannot find package "github.com/cockroachdb/cockroach/pkg/util/ctxgroup" in any of: /usr/local/go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup (from $GOROOT) /go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup (from $GOPATH) Makefile:756: recipe for target 'gotestdashi' failed make: *** [gotestdashi] Error 1 ```
test
util ctxgroup unknown failed under stress sha parameters tags goflags stress build found a failed test find pkg name node modules prune o type f name pb go exec rm github com cockroachdb cockroach vendor github com golang glog github com cockroachdb cockroach vendor golang org x tools go gcexportdata yes set e for dir in pkg acceptance cluster pkg build pkg ccl backupccl pkg ccl baseccl pkg ccl storageccl engineccl enginepbccl pkg ccl utilccl licenseccl pkg config pkg gossip pkg internal client pkg roachpb pkg rpc pkg server diagnosticspb pkg server serverpb pkg server status pkg settings cluster pkg sql distsqlrun pkg sql jobs pkg sql pgwire pgerror pkg sql sqlbase pkg sql stats pkg storage pkg storage engine enginepb pkg storage storagebase pkg ts tspb pkg util pkg util hlc pkg util log pkg util protoutil pkg util tracing do build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis plugin bin protoc gen gogoroach gogoroach out mgoogle api annotations proto github com grpc ecosystem grpc gateway third party googleapis google api mgoogle protobuf timestamp proto github com gogo protobuf types plugins grpc import prefix github com cockroachdb cockroach pkg pkg dir proto done touch bin cpp ccl protobuf sources checking for sys types h github com cockroachdb cockroach vendor github com golang lint yes sed i e gogoproto d go src github com cockroachdb cockroach c deps libroach protos roachpb data pb h go src github com cockroachdb cockroach c deps libroach protos roachpb internal pb h go src github com cockroachdb cockroach c deps libroach protos roachpb metadata pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb file registry pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb mvcc pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb rocksdb pb h go src github com cockroachdb cockroach c deps libroach protos util hlc legacy timestamp pb h go src github com cockroachdb cockroach c deps libroach protos util hlc timestamp pb h go src github com cockroachdb cockroach c deps libroach protos util unresolved addr pb h go src github com cockroachdb cockroach c deps libroach protos roachpb data pb cc go src github com cockroachdb cockroach c deps libroach protos roachpb internal pb cc go src github com cockroachdb cockroach c deps libroach protos roachpb metadata pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb file registry pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb mvcc pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb rocksdb pb cc go src github com cockroachdb cockroach c deps libroach protos util hlc legacy timestamp pb cc go src github com cockroachdb cockroach c deps libroach protos util hlc timestamp pb cc go src github com cockroachdb cockroach c deps libroach protos util unresolved addr pb cc github com cockroachdb cockroach vendor github com golang protobuf protoc gen go plugin github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway httprule github com cockroachdb cockroach vendor github com jteeuwen go bindata touch bin cpp protobuf sources github com cockroachdb cockroach vendor github com kisielk errcheck internal errcheck checking for sys stat h github com cockroachdb cockroach vendor github com golang protobuf protoc gen go generator github com cockroachdb cockroach vendor github com golang lint golint github com cockroachdb cockroach vendor golang org x tools cover github com cockroachdb cockroach vendor github com jteeuwen go bindata go bindata github com cockroachdb cockroach vendor github com kisielk errcheck yes github com cockroachdb cockroach vendor github com mattn goveralls checking for stdlib h github com cockroachdb cockroach vendor github com misspell cmd misspell scanning dependencies of target roach building cxx object cmakefiles roach dir batch cc o yes building cxx object cmakefiles roach dir comparator cc o building cxx object cmakefiles roach dir encoding cc o building cxx object cmakefiles roach dir chunked buffer cc o building cxx object cmakefiles roach dir db cc o checking for string h yes checking for memory h github com cockroachdb cockroach vendor github com mibk dupl suffixtree yes github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway descriptor github com cockroachdb cockroach vendor github com mibk dupl syntax checking for strings h github com cockroachdb cockroach vendor github com mibk dupl syntax golang yes github com cockroachdb cockroach vendor github com mibk dupl job github com cockroachdb cockroach vendor github com mibk dupl output checking for inttypes h github com cockroachdb cockroach vendor github com wadey gocovmerge github com cockroachdb cockroach vendor github com mibk dupl github com cockroachdb cockroach vendor github com golang dep yes github com cockroachdb cockroach vendor github com golang dep internal feedback github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway generator checking for stdint h github com cockroachdb cockroach vendor golang org x perf internal stats github com cockroachdb cockroach vendor golang org x perf storage benchfmt github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway gengateway github com cockroachdb cockroach vendor golang org x tools imports yes checking for unistd h github com cockroachdb cockroach vendor golang org x perf benchstat github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway github com cockroachdb cockroach vendor github com golang dep internal importers base yes github com cockroachdb cockroach vendor golang org x perf cmd benchstat github com cockroachdb cockroach vendor github com golang dep internal importers glide github com cockroachdb cockroach vendor github com golang dep internal importers glock github com cockroachdb cockroach vendor github com golang dep internal importers godep checking whether byte ordering is bigendian github com cockroachdb cockroach vendor github com golang dep internal importers govend github com cockroachdb cockroach vendor github com golang dep internal importers govendor github com cockroachdb cockroach vendor github com golang dep internal importers gvt github com cockroachdb cockroach vendor github com golang dep internal importers vndr github com cockroachdb cockroach vendor golang org x tools cmd goyacc go internal gccgoimporter go internal gcimporter go internal srcimporter github com cockroachdb cockroach vendor github com golang dep internal importers github com cockroachdb cockroach vendor golang org x tools cmd goimports github com cockroachdb cockroach vendor github com golang dep cmd dep no checking size of void go importer github com cockroachdb cockroach vendor golang org x tools cmd stringer checking size of int checking size of long checking size of long long checking size of intmax t checking build system type building cxx object cmakefiles roach dir engine cc o pc linux gnu checking host system type pc linux gnu checking whether pause instruction is compilable yes checking for ar ar checking malloc h usability yes checking malloc h presence yes checking for malloc h yes checking whether malloc usable size definition can use const argument no building cxx object cmakefiles roach dir merge cc o checking for library containing log building cxx object cmakefiles roach dir mvcc cc o lm checking whether attribute syntax is compilable yes checking whether compiler supports fvisibility hidden building cxx object cmakefiles roach dir options cc o yes checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether tls model attribute is compilable touch bin bootstrap building cxx object cmakefiles roach dir snapshot cc o yes checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether alloc size attribute is compilable no checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether format gnu printf attribute is compilable no checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether format printf attribute is compilable yes checking for a bsd compatible install usr bin install c checking for ranlib ranlib checking for ld usr bin ld checking for autoconf usr bin autoconf checking for memalign yes checking for valloc yes checking whether compiler supports yes checking whether compiler supports funroll loops yes checking unwind h usability yes checking unwind h presence yes checking for unwind h yes checking for unwind backtrace in lgcc yes checking configured backtracing method libgcc checking for sbrk yes checking whether utrace is compilable no checking whether valgrind is compilable no checking whether a program using builtin unreachable is compilable yes checking whether a program using builtin ffsl is compilable yes checking lg page missing version file and unable to generate it creating bogus version building cxx object cmakefiles roach dir protos roachpb data pb cc o checking pthread h usability yes checking pthread h presence yes checking for pthread h yes checking for pthread create in lpthread yes checking whether pthread atfork is compilable yes checking for library containing clock gettime none required checking whether clock gettime clock monotonic coarse is compilable yes checking whether clock gettime clock monotonic is compilable yes checking whether mach absolute time is compilable no checking whether compiler supports werror yes checking whether syscall is compilable yes checking for secure getenv yes checking for issetugid no checking for malloc thread cleanup building cxx object cmakefiles roach dir protos roachpb internal pb cc o no checking for pthread mutex init calloc cb no checking for tls yes checking whether atomics is compilable building cxx object cmakefiles roach dir protos roachpb metadata pb cc o yes checking whether atomic is compilable no checking whether darwin osatomic is compilable no checking whether madvise is compilable yes checking whether madvise madv free is compilable building cxx object cmakefiles roach dir protos storage engine enginepb mvcc pb cc o no checking whether madvise madv dontneed is compilable building cxx object cmakefiles roach dir protos storage engine enginepb pb cc o yes checking whether madvise madv hugepage is compilable building cxx object cmakefiles roach dir protos storage engine enginepb file registry pb cc o yes checking whether to force bit sync add sub and fetch no checking whether to force bit sync add sub and fetch no checking for builtin clz yes checking whether darwin os unfair lock is compilable no checking whether darwin osspin is compilable no checking whether glibc malloc hook is compilable yes checking whether glibc memalign hook is compilable yes checking whether pthreads adaptive mutexes is compilable yes checking for stdbool h that conforms to yes checking for bool yes configure creating config status config status creating makefile config status creating jemalloc pc config status creating doc html xsl config status creating doc manpages xsl config status creating doc jemalloc xml config status creating include jemalloc jemalloc macros h config status creating include jemalloc jemalloc protos h building cxx object cmakefiles roach dir protos storage engine enginepb rocksdb pb cc o config status creating include jemalloc jemalloc typedefs h config status creating include jemalloc internal jemalloc internal h config status creating test test sh config status creating test include test jemalloc test h config status creating config stamp config status creating bin jemalloc config config status creating bin jemalloc sh config status creating bin jeprof config status creating include jemalloc jemalloc defs h config status creating include jemalloc internal jemalloc internal defs h config status creating test include test jemalloc test defs h config status executing include jemalloc internal private namespace h commands config status executing include jemalloc internal private unnamespace h commands config status executing include jemalloc internal public symbols txt commands config status executing include jemalloc internal public namespace h commands building cxx object cmakefiles roach dir protos util hlc legacy timestamp pb cc o config status executing include jemalloc internal public unnamespace h commands building cxx object cmakefiles roach dir protos util hlc timestamp pb cc o config status executing include jemalloc internal size classes h commands building cxx object cmakefiles roach dir protos util unresolved addr pb cc o config status executing include jemalloc jemalloc protos jet h commands config status executing include jemalloc jemalloc rename h commands config status executing include jemalloc jemalloc mangle h commands set euo pipefail ret cd pkg sql parser gen goyacc p sql o sql go tmp sql y if expr ret conflicts dev null then echo ret exit fi config status executing include jemalloc jemalloc mangle jet h commands config status executing include jemalloc jemalloc h commands jemalloc version library revision config enable prof cc cc configure cflags std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops specified cflags extra cflags cppflags d gnu source d reentrant ldflags extra ldflags libs lm lgcc lm lpthread rpath extra xsltproc false xslroot prefix usr local bindir usr local bin datadir usr local share includedir usr local include libdir usr local lib mandir usr local share man srcroot go src github com cockroachdb cockroach c deps jemalloc abs srcroot go src github com cockroachdb cockroach c deps jemalloc objroot abs objroot go native pc linux gnu jemalloc jemalloc prefix jemalloc private namespace je install suffix malloc conf autogen cc silence debug code coverage stats prof prof libunwind prof libgcc prof gcc tcache thp fill utrace valgrind xmalloc munmap lazy lock tls cache oblivious find pkg name node modules prune o type f name pb gw go exec rm build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis grpc gateway out logtostderr true request context true pkg pkg server serverpb admin proto pkg server serverpb status proto pkg server serverpb authentication proto build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis grpc gateway out logtostderr true request context true pkg pkg ts tspb timeseries proto sed i import d pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s golang org x net context context g pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go gofmt s w pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go goimports w pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go linking cxx static library libroach a sed i e s import fmt math github com cockroachdb cockroach pkg fmt math g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go built target roach scanning dependencies of target roachccl building cxx object cmakefiles roachccl dir ccl db cc o building cxx object cmakefiles roachccl dir protosccl ccl storageccl engineccl enginepbccl key registry pb cc o building cxx object cmakefiles roachccl dir protosccl ccl baseccl encryption options pb cc o building cxx object cmakefiles roachccl dir ccl key manager cc o touch bin gw protobuf sources sed i e s cockroachdb cockroach pkg etcd coreos g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s github com cockroachdb cockroach pkg bytes encoding binary errors fmt io math github com google golang org g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s golang org x net context context g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go gofmt s w pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go touch bin go protobuf sources cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src arena o go src github com cockroachdb cockroach c deps jemalloc src arena c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src jemalloc o go src github com cockroachdb cockroach c deps jemalloc src jemalloc c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src atomic o go src github com cockroachdb cockroach c deps jemalloc src atomic c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src base o go src github com cockroachdb cockroach c deps jemalloc src base c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src bitmap o go src github com cockroachdb cockroach c deps jemalloc src bitmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk o go src github com cockroachdb cockroach c deps jemalloc src chunk c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk dss o go src github com cockroachdb cockroach c deps jemalloc src chunk dss c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk mmap o go src github com cockroachdb cockroach c deps jemalloc src chunk mmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ckh o go src github com cockroachdb cockroach c deps jemalloc src ckh c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ctl o go src github com cockroachdb cockroach c deps jemalloc src ctl c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src extent o go src github com cockroachdb cockroach c deps jemalloc src extent c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src hash o go src github com cockroachdb cockroach c deps jemalloc src hash c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidde echo code generated by goyacc do not edit echo generated file do not edit cat pkg sql parser gen sql go tmp sed e s const const lex g pkg sql parser sql go echo code generated by make do not edit echo generated file do not edit echo echo package lex echo grep const pkg sql parser gen sql go tmp pkg sql lex tokens go go run pkg cmd docgen main funcs go functions docs generated sql quiet linking cxx static library libroachccl a built target roachccl n funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src huge o go src github com cockroachdb cockroach c deps jemalloc src huge c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mb o go src github com cockroachdb cockroach c deps jemalloc src mb c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mutex o go src github com cockroachdb cockroach c deps jemalloc src mutex c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src nstime o go src github com cockroachdb cockroach c deps jemalloc src nstime c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src pages o go src github com cockroachdb cockroach c deps jemalloc src pages c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prng o go src github com cockroachdb cockroach c deps jemalloc src prng c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prof o go src github com cockroachdb cockroach c deps jemalloc src prof c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src quarantine o go src github com cockroachdb cockroach c deps jemalloc src quarantine c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src rtree o go src github com cockroachdb cockroach c deps jemalloc src rtree c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src stats o go src github com cockroachdb cockroach c deps jemalloc src stats c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src spin o go src github com cockroachdb cockroach c deps jemalloc src spin c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tcache o go src github com cockroachdb cockroach c deps jemalloc src tcache c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ticker o go src github com cockroachdb cockroach c deps jemalloc src ticker c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tsd o go src github com cockroachdb cockroach c deps jemalloc src tsd c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src util o go src github com cockroachdb cockroach c deps jemalloc src util c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src witness o go src github com cockroachdb cockroach c deps jemalloc src witness c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src jemalloc pic o go src github com cockroachdb cockroach c deps jemalloc src jemalloc c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src arena pic o go src github com cockroachdb cockroach c deps jemalloc src arena c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src atomic pic o go src github com cockroachdb cockroach c deps jemalloc src atomic c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src base pic o go src github com cockroachdb cockroach c deps jemalloc src base c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src bitmap pic o go src github com cockroachdb cockroach c deps jemalloc src bitmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk pic o go src github com cockroachdb cockroach c deps jemalloc src chunk c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk dss pic o go src github com cockroachdb cockroach c deps jemalloc src chunk dss c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk mmap pic o go src github com cockroachdb cockroach c deps jemalloc src chunk mmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ckh pic o go src github com cockroachdb cockroach c deps jemalloc src ckh c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ctl pic o go src github com cockroachdb cockroach c deps jemalloc src ctl c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src extent pic o go src github com cockroachdb cockroach c deps jemalloc src extent c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src hash pic o go src github com cockroachdb cockroach c deps jemalloc src hash c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src huge pic o go src github com cockroachdb cockroach c deps jemalloc src huge c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mb pic o go src github com cockroachdb cockroach c deps jemalloc src mb c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mutex pic o go src github com cockroachdb cockroach c deps jemalloc src mutex c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src nstime pic o go src github com cockroachdb cockroach c deps jemalloc src nstime c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src pages pic o go src github com cockroachdb cockroach c deps jemalloc src pages c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prng pic o go src github com cockroachdb cockroach c deps jemalloc src prng c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prof pic o go src github com cockroachdb cockroach c deps jemalloc src prof c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src quarantine pic o go src github com cockroachdb cockroach c deps jemalloc src quarantine c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go sar u modifier ignored since d is the default see u ar u modifier ignored since d is the default see u rc github com cockroachdb cockroach c deps jemalloc include iinclude o src rtree pic o go src github com cockroachdb cockroach c deps jemalloc src rtree c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src stats pic o go src github com cockroachdb cockroach c deps jemalloc src stats c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src spin pic o go src github com cockroachdb cockroach c deps jemalloc src spin c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tcache pic o go src github com cockroachdb cockroach c deps jemalloc src tcache c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ticker pic o go src github com cockroachdb cockroach c deps jemalloc src ticker c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tsd pic o go src github com cockroachdb cockroach c deps jemalloc src tsd c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src util pic o go src github com cockroachdb cockroach c deps jemalloc src util c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src witness pic o go src github com cockroachdb cockroach c deps jemalloc src witness c ar crus lib libjemalloc a src jemalloc o src arena o src atomic o src base o src bitmap o src chunk o src chunk dss o src chunk mmap o src ckh o src ctl o src extent o src hash o src huge o src mb o src mutex o src nstime o src pages o src prng o src prof o src quarantine o src rtree o src stats o src spin o src tcache o src ticker o src tsd o src util o src witness o ar crus lib libjemalloc pic a src jemalloc pic o src arena pic o src atomic pic o src base pic o src bitmap pic o src chunk pic o src chunk dss pic o src chunk mmap pic o src ckh pic o src ctl pic o src extent pic o src hash pic o src huge pic o src mb pic o src mutex pic o src nstime pic o src pages pic o src prng pic o src prof pic o src quarantine pic o src rtree pic o src stats pic o src spin pic o src tcache pic o src ticker pic o src tsd pic o src util pic o src witness pic o built target build version touch bin docgen functions scanning dependencies of target rocksdb building cxx object cmakefiles rocksdb dir cache clock cache cc o building cxx object cmakefiles rocksdb dir cache lru cache cc o building cxx object cmakefiles rocksdb dir cache sharded cache cc o building cxx object cmakefiles rocksdb dir db c cc o building cxx object cmakefiles rocksdb dir db builder cc o building cxx object cmakefiles rocksdb dir db compaction cc o building cxx object cmakefiles rocksdb dir db column family cc o building cxx object cmakefiles rocksdb dir db compacted db impl cc o building cxx object cmakefiles rocksdb dir db compaction iterator cc o building cxx object cmakefiles rocksdb dir db compaction job cc o building cxx object cmakefiles rocksdb dir db compaction picker cc o building cxx object cmakefiles rocksdb dir db compaction picker universal cc o building cxx object cmakefiles rocksdb dir db convenience cc o building cxx object cmakefiles rocksdb dir db db filesnapshot cc o building cxx object cmakefiles rocksdb dir db db impl cc o building cxx object cmakefiles rocksdb dir db db impl write cc o building cxx object cmakefiles rocksdb dir db db impl compaction flush cc o building cxx object cmakefiles rocksdb dir db db impl files cc o building cxx object cmakefiles rocksdb dir db db impl open cc o building cxx object cmakefiles rocksdb dir db db impl debug cc o building cxx object cmakefiles rocksdb dir db db impl experimental cc o building cxx object cmakefiles rocksdb dir db db impl readonly cc o building cxx object cmakefiles rocksdb dir db db info dumper cc o building cxx object cmakefiles rocksdb dir db db iter cc o building cxx object cmakefiles rocksdb dir db dbformat cc o building cxx object cmakefiles rocksdb dir db event helpers cc o building cxx object cmakefiles rocksdb dir db experimental cc o building cxx object cmakefiles rocksdb dir db external sst file ingestion job cc o building cxx object cmakefiles rocksdb dir db file indexer cc o building cxx object cmakefiles rocksdb dir db flush job cc o building cxx object cmakefiles rocksdb dir db flush scheduler cc o building cxx object cmakefiles rocksdb dir db forward iterator cc o building cxx object cmakefiles rocksdb dir db internal stats cc o building cxx object cmakefiles rocksdb dir db log reader cc o building cxx object cmakefiles rocksdb dir db log writer cc o building cxx object cmakefiles rocksdb dir db malloc stats cc o building cxx object cmakefiles rocksdb dir db managed iterator cc o building cxx object cmakefiles rocksdb dir db memtable cc o building cxx object cmakefiles rocksdb dir db memtable list cc o building cxx object cmakefiles rocksdb dir db merge helper cc o building cxx object cmakefiles rocksdb dir db range del aggregator cc o building cxx object cmakefiles rocksdb dir db repair cc o building cxx object cmakefiles rocksdb dir db table cache cc o building cxx object cmakefiles rocksdb dir db table properties collector cc o building cxx object cmakefiles rocksdb dir db transaction log impl cc o building cxx object cmakefiles rocksdb dir db version builder cc o building cxx object cmakefiles rocksdb dir db version edit cc o building cxx object cmakefiles rocksdb dir db version set cc o building cxx object cmakefiles rocksdb dir db wal manager cc o building cxx object cmakefiles rocksdb dir db write batch cc o building cxx object cmakefiles rocksdb dir db write thread cc o building cxx object cmakefiles rocksdb dir env env cc o building cxx object cmakefiles rocksdb dir env env encryption cc o building cxx object cmakefiles rocksdb dir env mock env cc o building cxx object cmakefiles rocksdb dir memtable alloc tracker cc o building cxx object cmakefiles rocksdb dir memtable hash cuckoo rep cc o building cxx object cmakefiles rocksdb dir memtable hash linklist rep cc o building cxx object cmakefiles rocksdb dir memtable hash skiplist rep cc o building cxx object cmakefiles rocksdb dir memtable skiplistrep cc o building cxx object cmakefiles rocksdb dir memtable vectorrep cc o building cxx object cmakefiles rocksdb dir memtable write buffer manager cc o building cxx object cmakefiles rocksdb dir monitoring histogram cc o building cxx object cmakefiles rocksdb dir monitoring instrumented mutex cc o building cxx object cmakefiles rocksdb dir monitoring iostats context cc o building cxx object cmakefiles rocksdb dir monitoring perf context cc o building cxx object cmakefiles rocksdb dir monitoring perf level cc o building cxx object cmakefiles rocksdb dir monitoring statistics cc o building cxx object cmakefiles rocksdb dir monitoring thread status updater cc o building cxx object cmakefiles rocksdb dir monitoring thread status util cc o building cxx object cmakefiles rocksdb dir monitoring thread status util debug cc o building cxx object cmakefiles rocksdb dir options cf options cc o building cxx object cmakefiles rocksdb dir options db options cc o building cxx object cmakefiles rocksdb dir options options cc o building cxx object cmakefiles rocksdb dir options options helper cc o building cxx object cmakefiles rocksdb dir options options parser cc o building cxx object cmakefiles rocksdb dir table adaptive table factory cc o building cxx object cmakefiles rocksdb dir table block cc o building cxx object cmakefiles rocksdb dir table block based filter block cc o building cxx object cmakefiles rocksdb dir table block based table builder cc o building cxx object cmakefiles rocksdb dir table block based table factory cc o building cxx object cmakefiles rocksdb dir table block based table reader cc o building cxx object cmakefiles rocksdb dir table block builder cc o building cxx object cmakefiles rocksdb dir table block prefix index cc o building cxx object cmakefiles rocksdb dir table bloom block cc o building cxx object cmakefiles rocksdb dir table cuckoo table builder cc o building cxx object cmakefiles rocksdb dir table cuckoo table factory cc o building cxx object cmakefiles rocksdb dir table cuckoo table reader cc o building cxx object cmakefiles rocksdb dir table format cc o building cxx object cmakefiles rocksdb dir table full filter block cc o building cxx object cmakefiles rocksdb dir table get context cc o building cxx object cmakefiles rocksdb dir table index builder cc o building cxx object cmakefiles rocksdb dir table iterator cc o building cxx object cmakefiles rocksdb dir table merging iterator cc o building cxx object cmakefiles rocksdb dir table meta blocks cc o building cxx object cmakefiles rocksdb dir table partitioned filter block cc o building cxx object cmakefiles rocksdb dir table persistent cache helper cc o building cxx object cmakefiles rocksdb dir table plain table builder cc o building cxx object cmakefiles rocksdb dir table plain table factory cc o building cxx object cmakefiles rocksdb dir table plain table index cc o building cxx object cmakefiles rocksdb dir table plain table key coding cc o building cxx object cmakefiles rocksdb dir table plain table reader cc o building cxx object cmakefiles rocksdb dir table sst file writer cc o building cxx object cmakefiles rocksdb dir table table properties cc o building cxx object cmakefiles rocksdb dir table two level iterator cc o building cxx object cmakefiles rocksdb dir tools db bench tool cc o building cxx object cmakefiles rocksdb dir tools dump db dump tool cc o building cxx object cmakefiles rocksdb dir tools ldb cmd cc o building cxx object cmakefiles rocksdb dir tools sst dump tool cc o building cxx object cmakefiles rocksdb dir util arena cc o building cxx object cmakefiles rocksdb dir util auto roll logger cc o building cxx object cmakefiles rocksdb dir util bloom cc o building cxx object cmakefiles rocksdb dir util coding cc o building cxx object cmakefiles rocksdb dir util comparator cc o building cxx object cmakefiles rocksdb dir util concurrent arena cc o building cxx object cmakefiles rocksdb dir util cc o building cxx object cmakefiles rocksdb dir util delete scheduler cc o building cxx object cmakefiles rocksdb dir util dynamic bloom cc o building cxx object cmakefiles rocksdb dir util event logger cc o building cxx object cmakefiles rocksdb dir util file reader writer cc o building cxx object cmakefiles rocksdb dir util file util cc o building cxx object cmakefiles rocksdb dir util filename cc o building cxx object cmakefiles rocksdb dir util hash cc o building cxx object cmakefiles rocksdb dir util log buffer cc o building cxx object cmakefiles rocksdb dir util random cc o building cxx object cmakefiles rocksdb dir util rate limiter cc o building cxx object cmakefiles rocksdb dir util sst file manager impl cc o building cxx object cmakefiles rocksdb dir util status cc o building cxx object cmakefiles rocksdb dir util sync point cc o building cxx object cmakefiles rocksdb dir util testutil cc o building cxx object cmakefiles rocksdb dir util thread local cc o building cxx object cmakefiles rocksdb dir util threadpool imp cc o building cxx object cmakefiles rocksdb dir util transaction test util cc o building cxx object cmakefiles rocksdb dir utilities backupable backupable db cc o building cxx object cmakefiles rocksdb dir utilities blob db blob db cc o building cxx object cmakefiles rocksdb dir utilities blob db blob db impl cc o building cxx object cmakefiles rocksdb dir utilities blob db blob dump tool cc o building cxx object cmakefiles rocksdb dir utilities blob db blob file cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log reader cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log writer cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log format cc o building cxx object cmakefiles rocksdb dir utilities blob db ttl extractor cc o building cxx object cmakefiles rocksdb dir utilities checkpoint checkpoint impl cc o building cxx object cmakefiles rocksdb dir utilities col buf decoder cc o building cxx object cmakefiles rocksdb dir utilities col buf encoder cc o building cxx object cmakefiles rocksdb dir utilities column aware encoding util cc o building cxx object cmakefiles rocksdb dir utilities date tiered date tiered db impl cc o building cxx object cmakefiles rocksdb dir utilities debug cc o building cxx object cmakefiles rocksdb dir utilities document document db cc o building cxx object cmakefiles rocksdb dir utilities document json document cc o building cxx object cmakefiles rocksdb dir utilities env timed cc o building cxx object cmakefiles rocksdb dir utilities geodb geodb impl cc o building cxx object cmakefiles rocksdb dir utilities memory memory util cc o building cxx object cmakefiles rocksdb dir utilities merge operators cc o building cxx object cmakefiles rocksdb dir utilities options options util cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier file cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier metadata cc o building cxx object cmakefiles rocksdb dir utilities persistent cache volatile tier impl cc o building cxx object cmakefiles rocksdb dir utilities redis redis lists cc o building cxx object cmakefiles rocksdb dir utilities simulator cache sim cache cc o building cxx object cmakefiles rocksdb dir utilities spatialdb spatial db cc o building cxx object cmakefiles rocksdb dir utilities transactions optimistic transaction db impl cc o building cxx object cmakefiles rocksdb dir utilities transactions optimistic transaction cc o building cxx object cmakefiles rocksdb dir utilities transactions pessimistic transaction cc o building cxx object cmakefiles rocksdb dir utilities transactions pessimistic transaction db cc o building cxx object cmakefiles rocksdb dir utilities transactions snapshot checker cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction base cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction lock mgr cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction util cc o building cxx object cmakefiles rocksdb dir utilities transactions write prepared txn cc o building cxx object cmakefiles rocksdb dir utilities ttl db ttl impl cc o building cxx object cmakefiles rocksdb dir utilities write batch with index write batch with index cc o building cxx object cmakefiles rocksdb dir utilities write batch with index write batch with index internal cc o building cxx object cmakefiles rocksdb dir port port posix cc o building cxx object cmakefiles rocksdb dir env env posix cc o building cxx object cmakefiles rocksdb dir env io posix cc o linking cxx static library librocksdb a built target rocksdb go test v tags make pc linux gnu ldflags x github com cockroachdb cockroach pkg build typ development x github com cockroachdb cockroach pkg build tag dirty x github com cockroachdb cockroach pkg build utctime x github com cockroachdb cockroach pkg build rev x github com cockroachdb cockroach pkg build cgotargettriple pc linux gnu i github com cockroachdb cockroach pkg util ctxgroup can t load package package github com cockroachdb cockroach pkg util ctxgroup cannot find package github com cockroachdb cockroach pkg util ctxgroup in any of usr local go src github com cockroachdb cockroach pkg util ctxgroup from goroot go src github com cockroachdb cockroach pkg util ctxgroup from gopath makefile recipe for target gotestdashi failed make error
1
229,012
17,497,690,164
IssuesEvent
2021-08-10 04:21:56
work-teams/software-quality
https://api.github.com/repos/work-teams/software-quality
opened
Actualizar README.md con los informes
documentation
- Agregar el apartado "About" sobre el proyecto. - Mejorar el diseño de los integrantes. - Agregar sección instalación del proyecto. - Agregar sección manuales de usuario (vendedor, administrador). - Agregar sección métricas del proyecto.
1.0
Actualizar README.md con los informes - - Agregar el apartado "About" sobre el proyecto. - Mejorar el diseño de los integrantes. - Agregar sección instalación del proyecto. - Agregar sección manuales de usuario (vendedor, administrador). - Agregar sección métricas del proyecto.
non_test
actualizar readme md con los informes agregar el apartado about sobre el proyecto mejorar el diseño de los integrantes agregar sección instalación del proyecto agregar sección manuales de usuario vendedor administrador agregar sección métricas del proyecto
0
180,879
13,963,615,168
IssuesEvent
2020-10-25 14:53:04
pytest-dev/pytest
https://api.github.com/repos/pytest-dev/pytest
closed
testing: test_pdb_unittest_postmortem get stuck with pexpect
type: selftests
This causes the ubuntu-py37 CI job to get stuck and timeout after 6 hours. This job runs `tox -e py37-lsof-numpy-pexpect` but I can reproduce locally with `tox -e py38-pexpect` so it's not about the python version or lsof/numpy. When running just `tox -e py-pexpect -- testing/test_debugging.py` it does not happen; so it's only with some interaction with other tests. The pexpect timeout evidentially doesn't kick in. Don't have a useful traceback. Bisected to 69419cb70007514e5061f3b8d359b7da9dff82dd (new pytester fixture) - cc @nicoddemus. That's what I had time to investigate for now.
1.0
testing: test_pdb_unittest_postmortem get stuck with pexpect - This causes the ubuntu-py37 CI job to get stuck and timeout after 6 hours. This job runs `tox -e py37-lsof-numpy-pexpect` but I can reproduce locally with `tox -e py38-pexpect` so it's not about the python version or lsof/numpy. When running just `tox -e py-pexpect -- testing/test_debugging.py` it does not happen; so it's only with some interaction with other tests. The pexpect timeout evidentially doesn't kick in. Don't have a useful traceback. Bisected to 69419cb70007514e5061f3b8d359b7da9dff82dd (new pytester fixture) - cc @nicoddemus. That's what I had time to investigate for now.
test
testing test pdb unittest postmortem get stuck with pexpect this causes the ubuntu ci job to get stuck and timeout after hours this job runs tox e lsof numpy pexpect but i can reproduce locally with tox e pexpect so it s not about the python version or lsof numpy when running just tox e py pexpect testing test debugging py it does not happen so it s only with some interaction with other tests the pexpect timeout evidentially doesn t kick in don t have a useful traceback bisected to new pytester fixture cc nicoddemus that s what i had time to investigate for now
1
31,182
4,697,200,507
IssuesEvent
2016-10-12 08:31:56
MachoThemes/newsmag-lite
https://api.github.com/repos/MachoThemes/newsmag-lite
closed
Join "Posts colum" + "Posts list vertical small" into one widget
enhancement needs testing tested
- [x] Epsilon toggle for "Featured article" - [x] Epsilon toggle for "Display the article's date"
2.0
Join "Posts colum" + "Posts list vertical small" into one widget - - [x] Epsilon toggle for "Featured article" - [x] Epsilon toggle for "Display the article's date"
test
join posts colum posts list vertical small into one widget epsilon toggle for featured article epsilon toggle for display the article s date
1
92,824
8,378,935,522
IssuesEvent
2018-10-06 19:18:49
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: ycsb/F/nodes=3 failed
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/bd9de99ff5569d9c21df027544b4fdb17ffc4a55 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stressrace instead of stress and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ make stress TESTS=ycsb/F/nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m' ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=946966&tab=buildLog ``` The test failed on master: test.go:500,cluster.go:556: mkdir /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20181006-946966/ycsb: no space left on device ```
1.0
roachtest: ycsb/F/nodes=3 failed - SHA: https://github.com/cockroachdb/cockroach/commits/bd9de99ff5569d9c21df027544b4fdb17ffc4a55 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stressrace instead of stress and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ make stress TESTS=ycsb/F/nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m' ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=946966&tab=buildLog ``` The test failed on master: test.go:500,cluster.go:556: mkdir /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20181006-946966/ycsb: no space left on device ```
test
roachtest ycsb f nodes failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stressrace instead of stress and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach make stress tests ycsb f nodes pkg roachtest testtimeout stressflags stderr false maxtime timeout failed test the test failed on master test go cluster go mkdir home agent work go src github com cockroachdb cockroach artifacts ycsb no space left on device
1
107,616
9,218,987,345
IssuesEvent
2019-03-11 14:33:48
pravega/pravega
https://api.github.com/repos/pravega/pravega
opened
Split testZKSessionExpiry into two separate tests
area/controller area/testing version/0.5.0
**Problem description** In PR #3387, `testZKSessionExpiry` has been extended to test 2 different scenarios (with/without closing the ZK ckient). This makes the test to be too long for being a unit test. **Problem location** `ZKControllerServiceMainTest` **Suggestions for an improvement** Split the test into the 2 different scenarios tested.
1.0
Split testZKSessionExpiry into two separate tests - **Problem description** In PR #3387, `testZKSessionExpiry` has been extended to test 2 different scenarios (with/without closing the ZK ckient). This makes the test to be too long for being a unit test. **Problem location** `ZKControllerServiceMainTest` **Suggestions for an improvement** Split the test into the 2 different scenarios tested.
test
split testzksessionexpiry into two separate tests problem description in pr testzksessionexpiry has been extended to test different scenarios with without closing the zk ckient this makes the test to be too long for being a unit test problem location zkcontrollerservicemaintest suggestions for an improvement split the test into the different scenarios tested
1
25,926
3,974,143,690
IssuesEvent
2016-05-04 21:05:33
sfbrigade/stop-violence
https://api.github.com/repos/sfbrigade/stop-violence
opened
Design a Project Description page
design
We need a static page to describe what the project is about, who we are, and why we made it. Initial wireframe (showing req'd content only): https://github.com/sfbrigade/stop-violence/blob/master/design/proj_desc_wire.png
1.0
Design a Project Description page - We need a static page to describe what the project is about, who we are, and why we made it. Initial wireframe (showing req'd content only): https://github.com/sfbrigade/stop-violence/blob/master/design/proj_desc_wire.png
non_test
design a project description page we need a static page to describe what the project is about who we are and why we made it initial wireframe showing req d content only
0
265,713
23,192,050,175
IssuesEvent
2022-08-01 13:27:05
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Node3D::look_at_from_position broken in Godot 4 alpha 12
bug topic:core needs testing topic:3d
### Godot version v4.0.alpha12.official [2c11e6d9e] ### System information Xubuntu 22.04 ### Issue description I'm pretty sure set_scale(original_scale) at the end of Node3D::look_at_from_position resets the rotation of the node. I didn't re-compile the engine to verify, but I can reproduce the bug by reimplementing look_at_from_position using gdscript, and removing the set_scale fixes it. Maybe it was something introduced with commit 7acf6974792 ### Steps to reproduce Use look_at or look_at_from_position on a Node3D ### Minimal reproduction project _No response_
1.0
Node3D::look_at_from_position broken in Godot 4 alpha 12 - ### Godot version v4.0.alpha12.official [2c11e6d9e] ### System information Xubuntu 22.04 ### Issue description I'm pretty sure set_scale(original_scale) at the end of Node3D::look_at_from_position resets the rotation of the node. I didn't re-compile the engine to verify, but I can reproduce the bug by reimplementing look_at_from_position using gdscript, and removing the set_scale fixes it. Maybe it was something introduced with commit 7acf6974792 ### Steps to reproduce Use look_at or look_at_from_position on a Node3D ### Minimal reproduction project _No response_
test
look at from position broken in godot alpha godot version official system information xubuntu issue description i m pretty sure set scale original scale at the end of look at from position resets the rotation of the node i didn t re compile the engine to verify but i can reproduce the bug by reimplementing look at from position using gdscript and removing the set scale fixes it maybe it was something introduced with commit steps to reproduce use look at or look at from position on a minimal reproduction project no response
1
89,613
8,210,025,398
IssuesEvent
2018-09-04 09:26:27
elastic/beats
https://api.github.com/repos/elastic/beats
opened
Improve kubernetes integration tests
:Testing enhancement
Current implementation (based on localkube) is unreliable, we should invest some time in an alternative way to do Kubernetes integration tests. Probably using minikube or kubeadm (dind flavor)
1.0
Improve kubernetes integration tests - Current implementation (based on localkube) is unreliable, we should invest some time in an alternative way to do Kubernetes integration tests. Probably using minikube or kubeadm (dind flavor)
test
improve kubernetes integration tests current implementation based on localkube is unreliable we should invest some time in an alternative way to do kubernetes integration tests probably using minikube or kubeadm dind flavor
1
72,390
24,095,500,632
IssuesEvent
2022-09-19 18:21:42
idaholab/HERON
https://api.github.com/repos/idaholab/HERON
closed
[DEFECT] Workshop ARMA needs to be retrained.
defect
-------- Defect Description -------- Due to some recent changes to the ARMA, we need to retrain the ARMA pk file in the workshop. ``` Traceback (most recent call last): File "/Users/mcdodj/Documents/projects/HERON/src/main.py", line 119, in <module> sim.read_input(args.xml_input_file) # TODO expand to use arguments? File "/Users/mcdodj/Documents/projects/HERON/src/main.py", line 53, in read_input objects = input_loader.parse(inp, location, self.messageHandler) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/input_loader.py", line 104, in parse case.initialize(components, sources) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/Cases.py", line 668, in initialize src.checkValid(self, components, sources) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/Placeholders.py", line 199, in checkValid structure = hutils.get_synthhist_structure(self._target_file) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/_utils.py", line 124, in get_synthhist_structure meta = rom.writeXML().getRoot() File "/Users/mcdodj/Documents/projects/raven/ravenframework/Models/ROM.py", line 599, in writeXML engines[0].writeXML(xml) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ROMCollection.py", line 1604, in writeXML step.writeXML(newNode, targets, skip) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ROMCollection.py", line 977, in writeXML Segments.writeXML(self, writeTo, targets, skip) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ROMCollection.py", line 401, in writeXML rom.writeXML(newNode, targets, skip) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ARMA.py", line 1394, in writeXML armaNode.append(xmlUtils.newNode('std', text=arma.sigma)) File "/usr/local/Caskroom/miniconda/base/envs/raven_libraries/lib/python3.7/site-packages/statsmodels/base/wrapper.py", line 34, in __getattribute__ obj = getattr(results, attr) AttributeError: 'ARMAResults' object has no attribute 'sigma' ... Errors were encountered while running HERON. ``` ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
1.0
[DEFECT] Workshop ARMA needs to be retrained. - -------- Defect Description -------- Due to some recent changes to the ARMA, we need to retrain the ARMA pk file in the workshop. ``` Traceback (most recent call last): File "/Users/mcdodj/Documents/projects/HERON/src/main.py", line 119, in <module> sim.read_input(args.xml_input_file) # TODO expand to use arguments? File "/Users/mcdodj/Documents/projects/HERON/src/main.py", line 53, in read_input objects = input_loader.parse(inp, location, self.messageHandler) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/input_loader.py", line 104, in parse case.initialize(components, sources) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/Cases.py", line 668, in initialize src.checkValid(self, components, sources) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/Placeholders.py", line 199, in checkValid structure = hutils.get_synthhist_structure(self._target_file) File "/Users/mcdodj/Documents/projects/HERON/src/../../HERON/src/_utils.py", line 124, in get_synthhist_structure meta = rom.writeXML().getRoot() File "/Users/mcdodj/Documents/projects/raven/ravenframework/Models/ROM.py", line 599, in writeXML engines[0].writeXML(xml) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ROMCollection.py", line 1604, in writeXML step.writeXML(newNode, targets, skip) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ROMCollection.py", line 977, in writeXML Segments.writeXML(self, writeTo, targets, skip) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ROMCollection.py", line 401, in writeXML rom.writeXML(newNode, targets, skip) File "/Users/mcdodj/Documents/projects/raven/ravenframework/SupervisedLearning/ARMA.py", line 1394, in writeXML armaNode.append(xmlUtils.newNode('std', text=arma.sigma)) File "/usr/local/Caskroom/miniconda/base/envs/raven_libraries/lib/python3.7/site-packages/statsmodels/base/wrapper.py", line 34, in __getattribute__ obj = getattr(results, attr) AttributeError: 'ARMAResults' object has no attribute 'sigma' ... Errors were encountered while running HERON. ``` ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
non_test
workshop arma needs to be retrained defect description due to some recent changes to the arma we need to retrain the arma pk file in the workshop traceback most recent call last file users mcdodj documents projects heron src main py line in sim read input args xml input file todo expand to use arguments file users mcdodj documents projects heron src main py line in read input objects input loader parse inp location self messagehandler file users mcdodj documents projects heron src heron src input loader py line in parse case initialize components sources file users mcdodj documents projects heron src heron src cases py line in initialize src checkvalid self components sources file users mcdodj documents projects heron src heron src placeholders py line in checkvalid structure hutils get synthhist structure self target file file users mcdodj documents projects heron src heron src utils py line in get synthhist structure meta rom writexml getroot file users mcdodj documents projects raven ravenframework models rom py line in writexml engines writexml xml file users mcdodj documents projects raven ravenframework supervisedlearning romcollection py line in writexml step writexml newnode targets skip file users mcdodj documents projects raven ravenframework supervisedlearning romcollection py line in writexml segments writexml self writeto targets skip file users mcdodj documents projects raven ravenframework supervisedlearning romcollection py line in writexml rom writexml newnode targets skip file users mcdodj documents projects raven ravenframework supervisedlearning arma py line in writexml armanode append xmlutils newnode std text arma sigma file usr local caskroom miniconda base envs raven libraries lib site packages statsmodels base wrapper py line in getattribute obj getattr results attr attributeerror armaresults object has no attribute sigma errors were encountered while running heron for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
0
301,925
26,109,539,700
IssuesEvent
2022-12-27 17:36:08
nrwl/nx
https://api.github.com/repos/nrwl/nx
closed
e2e shared libraries with Cypress imports break Jest types in other libraries in VSCode
type: bug scope: testing tools
## Current Behavior When we have a shared library for e2e utilities that have a paths alias in `tsconfig.base.json`, the Jest types in `.spec.ts` unit tests in other libraries break in VSCode with `Property 'toEqual' does not exist on type 'Assertion'.` This has a large developer experience impact for VSCode users. From what I can see, this only impacts VSCode but I'm not familiar enough with VSCode TypeScript support to raise an issue there or even find if this is a known issue. This does not impact running Jest via Nx or `jest` directly. ## Expected Behavior e2e shared libraries can have `paths` entries in `tsconfig.base.json` and pull in dependencies that include Cypress types without causing Jest types to break in VSCode (or generally). ## Steps to Reproduce 1. Clone https://github.com/nrwl/nx-examples 2. Open `libs/products/home-page/src/lib/home-page/home-page.component.spec.ts` in VSCode and see the Jest types load correctly 3. Open `libs/shared/e2e-utils/src/lib/shared-e2e-utils.ts` and add `import { defineConfig } from 'cypress';` at the top of the file 4. Go back to `libs/products/home-page/src/lib/home-page/home-page.component.spec.ts` and see the Jest types are now broken 5. Installing and using other dependencies that pull in Cypress types also causes the issue. In our case, using `@badeball/cypress-cucumber-preprocessor` and having `import { defineStep } from '@badeball/cypress-cucumber-preprocessor';` at the top of `libs/shared/e2e-utils/src/lib/shared-e2e-utils.ts` also causes the issue. ## Additional Info Removing imports that bring in Cypress types from the shared library or removing the `paths` alias for the e2e library make the type errors in VSCode go away. Another workaround is to skip Jest global types and import the types directly `import {expect, jest, test} from '@jest/globals';` Another workaround that maintains global types that we applied is to create a `tsconfig.cypress.json`, copy all of the same `paths` from `tsconfig.base.json`, remove e2e library paths from `tsconfig.base.json`, and update the e2e shared and test libraries to extend the `tsconfig.cypress.json`. This is all a manual process to manage, requires additional work on top of the current generators to setup for new libraries, and I'm not sure it works with Nx affected analysis on changes to `tsconfig.cypress.json`, etc. Related issues https://github.com/nrwl/nx/issues/892 https://github.com/nrwl/nx/issues/816 https://github.com/cypress-io/cypress/issues/22059 https://github.com/cypress-io/cypress/issues/6156 https://github.com/microsoft/TypeScript/issues/22331 ### Failure Logs N/A ### Environment VSCode: `1.73.0` TypeScript: `typescript` Cypress: `10.7.0` (any `10.x`+ version, it seems) Nx: I've reproduced on `14.x` and `15.x`
1.0
e2e shared libraries with Cypress imports break Jest types in other libraries in VSCode - ## Current Behavior When we have a shared library for e2e utilities that have a paths alias in `tsconfig.base.json`, the Jest types in `.spec.ts` unit tests in other libraries break in VSCode with `Property 'toEqual' does not exist on type 'Assertion'.` This has a large developer experience impact for VSCode users. From what I can see, this only impacts VSCode but I'm not familiar enough with VSCode TypeScript support to raise an issue there or even find if this is a known issue. This does not impact running Jest via Nx or `jest` directly. ## Expected Behavior e2e shared libraries can have `paths` entries in `tsconfig.base.json` and pull in dependencies that include Cypress types without causing Jest types to break in VSCode (or generally). ## Steps to Reproduce 1. Clone https://github.com/nrwl/nx-examples 2. Open `libs/products/home-page/src/lib/home-page/home-page.component.spec.ts` in VSCode and see the Jest types load correctly 3. Open `libs/shared/e2e-utils/src/lib/shared-e2e-utils.ts` and add `import { defineConfig } from 'cypress';` at the top of the file 4. Go back to `libs/products/home-page/src/lib/home-page/home-page.component.spec.ts` and see the Jest types are now broken 5. Installing and using other dependencies that pull in Cypress types also causes the issue. In our case, using `@badeball/cypress-cucumber-preprocessor` and having `import { defineStep } from '@badeball/cypress-cucumber-preprocessor';` at the top of `libs/shared/e2e-utils/src/lib/shared-e2e-utils.ts` also causes the issue. ## Additional Info Removing imports that bring in Cypress types from the shared library or removing the `paths` alias for the e2e library make the type errors in VSCode go away. Another workaround is to skip Jest global types and import the types directly `import {expect, jest, test} from '@jest/globals';` Another workaround that maintains global types that we applied is to create a `tsconfig.cypress.json`, copy all of the same `paths` from `tsconfig.base.json`, remove e2e library paths from `tsconfig.base.json`, and update the e2e shared and test libraries to extend the `tsconfig.cypress.json`. This is all a manual process to manage, requires additional work on top of the current generators to setup for new libraries, and I'm not sure it works with Nx affected analysis on changes to `tsconfig.cypress.json`, etc. Related issues https://github.com/nrwl/nx/issues/892 https://github.com/nrwl/nx/issues/816 https://github.com/cypress-io/cypress/issues/22059 https://github.com/cypress-io/cypress/issues/6156 https://github.com/microsoft/TypeScript/issues/22331 ### Failure Logs N/A ### Environment VSCode: `1.73.0` TypeScript: `typescript` Cypress: `10.7.0` (any `10.x`+ version, it seems) Nx: I've reproduced on `14.x` and `15.x`
test
shared libraries with cypress imports break jest types in other libraries in vscode current behavior when we have a shared library for utilities that have a paths alias in tsconfig base json the jest types in spec ts unit tests in other libraries break in vscode with property toequal does not exist on type assertion this has a large developer experience impact for vscode users from what i can see this only impacts vscode but i m not familiar enough with vscode typescript support to raise an issue there or even find if this is a known issue this does not impact running jest via nx or jest directly expected behavior shared libraries can have paths entries in tsconfig base json and pull in dependencies that include cypress types without causing jest types to break in vscode or generally steps to reproduce clone open libs products home page src lib home page home page component spec ts in vscode and see the jest types load correctly open libs shared utils src lib shared utils ts and add import defineconfig from cypress at the top of the file go back to libs products home page src lib home page home page component spec ts and see the jest types are now broken installing and using other dependencies that pull in cypress types also causes the issue in our case using badeball cypress cucumber preprocessor and having import definestep from badeball cypress cucumber preprocessor at the top of libs shared utils src lib shared utils ts also causes the issue additional info removing imports that bring in cypress types from the shared library or removing the paths alias for the library make the type errors in vscode go away another workaround is to skip jest global types and import the types directly import expect jest test from jest globals another workaround that maintains global types that we applied is to create a tsconfig cypress json copy all of the same paths from tsconfig base json remove library paths from tsconfig base json and update the shared and test libraries to extend the tsconfig cypress json this is all a manual process to manage requires additional work on top of the current generators to setup for new libraries and i m not sure it works with nx affected analysis on changes to tsconfig cypress json etc related issues failure logs n a environment vscode typescript typescript cypress any x version it seems nx i ve reproduced on x and x
1
716,214
24,625,640,346
IssuesEvent
2022-10-16 13:36:34
AY2223S1-CS2103T-T14-3/tp
https://api.github.com/repos/AY2223S1-CS2103T-T14-3/tp
closed
Add feature where displayed record list is always sorted by date
enhancement priority.Medium
- Displayed record list of a patient will remain sorted by date, in descending order, after executing commands that manipulate it, e.g. `addR` and `deleteR`.
1.0
Add feature where displayed record list is always sorted by date - - Displayed record list of a patient will remain sorted by date, in descending order, after executing commands that manipulate it, e.g. `addR` and `deleteR`.
non_test
add feature where displayed record list is always sorted by date displayed record list of a patient will remain sorted by date in descending order after executing commands that manipulate it e g addr and deleter
0
16,868
9,541,603,030
IssuesEvent
2019-04-30 23:02:25
stanford-rc/fuse-migratefs
https://api.github.com/repos/stanford-rc/fuse-migratefs
closed
Optimize open(O_TRUNC)
performance
Avoid costly copyup operation when the destination file is to be entirely rewritten . For instance, on `truncate()`/`ftruncate()` with `length = 0`, we know that the file will be completely overwritten, so it's not necessary to copyup the old data from the lower layer,as it will be discarded after the copyup anyway,
True
Optimize open(O_TRUNC) - Avoid costly copyup operation when the destination file is to be entirely rewritten . For instance, on `truncate()`/`ftruncate()` with `length = 0`, we know that the file will be completely overwritten, so it's not necessary to copyup the old data from the lower layer,as it will be discarded after the copyup anyway,
non_test
optimize open o trunc avoid costly copyup operation when the destination file is to be entirely rewritten for instance on truncate ftruncate with length we know that the file will be completely overwritten so it s not necessary to copyup the old data from the lower layer as it will be discarded after the copyup anyway
0
196,605
22,442,162,394
IssuesEvent
2022-06-21 02:35:51
valdisiljuconoks/episerverless
https://api.github.com/repos/valdisiljuconoks/episerverless
closed
WS-2019-0064 (High) detected in handlebars-1.3.0.tgz - autoclosed
security vulnerability
## WS-2019-0064 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p> <p>Path to dependency file: episerverless/src/Web/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json,episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - istanbul-0.2.16.tgz - :x: **handlebars-1.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/valdisiljuconoks/episerverless/commits/b7f1357806194699d26d4407637fd7e74151d063">b7f1357806194699d26d4407637fd7e74151d063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server. <p>Publish Date: 2019-01-30 <p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0064 (High) detected in handlebars-1.3.0.tgz - autoclosed - ## WS-2019-0064 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p> <p>Path to dependency file: episerverless/src/Web/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json,episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - istanbul-0.2.16.tgz - :x: **handlebars-1.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/valdisiljuconoks/episerverless/commits/b7f1357806194699d26d4407637fd7e74151d063">b7f1357806194699d26d4407637fd7e74151d063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server. <p>Publish Date: 2019-01-30 <p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws high detected in handlebars tgz autoclosed ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file episerverless src web modules protected shell shell clientresources lib xstyle package json path to vulnerable library episerverless packages episerver cms ui content modules protected shell shell clientresources lib xstyle node modules handlebars package json episerverless packages episerver cms ui content modules protected shell shell clientresources lib xstyle node modules handlebars package json dependency hierarchy intern geezer tgz root library istanbul tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details versions of handlebars prior to are vulnerable to prototype pollution templates may alter an objects prototype thus allowing an attacker to execute arbitrary code on the server publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
304,481
26,280,682,732
IssuesEvent
2023-01-07 08:59:48
cuny-academic-commons/commons-in-a-box
https://api.github.com/repos/cuny-academic-commons/commons-in-a-box
closed
Member Settings > Types: Singular/Plural labels for new type not pre-filled
OpenLab Needs testing
When I create a new member type, the fields for the singular and plural labels are pre-filled with '[object Object]' rather than the new member type name. I can delete that and replace it with the correct label and successfully save the correct labels. It's only the initial state that's incorrect. Thanks! <img width="1140" alt="Screen Shot 2023-01-02 at 11 29 56 AM" src="https://user-images.githubusercontent.com/7189824/210260298-47fc3ecb-0419-4f79-b361-c3ab71f46b03.png">
1.0
Member Settings > Types: Singular/Plural labels for new type not pre-filled - When I create a new member type, the fields for the singular and plural labels are pre-filled with '[object Object]' rather than the new member type name. I can delete that and replace it with the correct label and successfully save the correct labels. It's only the initial state that's incorrect. Thanks! <img width="1140" alt="Screen Shot 2023-01-02 at 11 29 56 AM" src="https://user-images.githubusercontent.com/7189824/210260298-47fc3ecb-0419-4f79-b361-c3ab71f46b03.png">
test
member settings types singular plural labels for new type not pre filled when i create a new member type the fields for the singular and plural labels are pre filled with rather than the new member type name i can delete that and replace it with the correct label and successfully save the correct labels it s only the initial state that s incorrect thanks img width alt screen shot at am src
1
151,442
19,651,468,117
IssuesEvent
2022-01-10 07:46:48
Shai-Demo-Org/JS-Demo
https://api.github.com/repos/Shai-Demo-Org/JS-Demo
opened
CVE-2019-10744 (High) detected in multiple libraries
security vulnerability
## CVE-2019-10744 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-2.4.2.tgz</b>, <b>lodash-4.17.11.tgz</b>, <b>lodash-4.13.1.tgz</b></p></summary> <p> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/zaproxy/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - async-2.6.1.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.13.1.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz">https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/nyc/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - istanbul-lib-instrument-1.1.0-alpha.4.tgz - babel-types-6.11.1.tgz - :x: **lodash-4.13.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Shai-Demo-Org/JS-Demo/commit/dfb4f8933cba5269dad89ff4f3e25e4dc1939df7">dfb4f8933cba5269dad89ff4f3e25e4dc1939df7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-07-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744>CVE-2019-10744</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution: lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;lodash:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.11","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"async:2.6.1;lodash:4.17.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.13.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-lib-instrument:1.1.0-alpha.4;babel-types:6.11.1;lodash:4.13.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2019-10744","vulnerabilityDetails":"Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744","cvss3Severity":"high","cvss3Score":"9.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-10744 (High) detected in multiple libraries - ## CVE-2019-10744 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-2.4.2.tgz</b>, <b>lodash-4.17.11.tgz</b>, <b>lodash-4.13.1.tgz</b></p></summary> <p> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/zaproxy/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - async-2.6.1.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.13.1.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz">https://registry.npmjs.org/lodash/-/lodash-4.13.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/nyc/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - istanbul-lib-instrument-1.1.0-alpha.4.tgz - babel-types-6.11.1.tgz - :x: **lodash-4.13.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Shai-Demo-Org/JS-Demo/commit/dfb4f8933cba5269dad89ff4f3e25e4dc1939df7">dfb4f8933cba5269dad89ff4f3e25e4dc1939df7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-07-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744>CVE-2019-10744</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution: lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;lodash:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.11","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"async:2.6.1;lodash:4.17.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.13.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-lib-instrument:1.1.0-alpha.4;babel-types:6.11.1;lodash:4.13.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash-4.17.12, lodash-amd-4.17.12, lodash-es-4.17.12, lodash.defaultsdeep-4.6.1, lodash.merge- 4.6.2, lodash.mergewith-4.6.2, lodash.template-4.5.0","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2019-10744","vulnerabilityDetails":"Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744","cvss3Severity":"high","cvss3Score":"9.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file package json path to vulnerable library node modules zaproxy node modules lodash package json dependency hierarchy zaproxy tgz root library x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy async tgz root library x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules nyc node modules lodash package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz istanbul lib instrument alpha tgz babel types tgz x lodash tgz vulnerable library found in head commit a href vulnerability details versions of lodash lower than are vulnerable to prototype pollution the function defaultsdeep could be tricked into adding or modifying properties of object prototype using a constructor payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash lodash amd lodash es lodash defaultsdeep lodash merge lodash mergewith lodash template isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree zaproxy lodash isminimumfixversionavailable true minimumfixversion lodash lodash amd lodash es lodash defaultsdeep lodash merge lodash mergewith lodash template isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency true dependencytree async lodash isminimumfixversionavailable true minimumfixversion lodash lodash amd lodash es lodash defaultsdeep lodash merge lodash mergewith lodash template isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency true dependencytree grunt if grunt contrib nodeunit nodeunit tap nyc istanbul lib instrument alpha babel types lodash isminimumfixversionavailable true minimumfixversion lodash lodash amd lodash es lodash defaultsdeep lodash merge lodash mergewith lodash template isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails versions of lodash lower than are vulnerable to prototype pollution the function defaultsdeep could be tricked into adding or modifying properties of object prototype using a constructor payload vulnerabilityurl
0
682,774
23,357,042,999
IssuesEvent
2022-08-10 08:23:03
Naisten-Linja/naisten-linja-services
https://api.github.com/repos/Naisten-Linja/naisten-linja-services
closed
Allow staff to unassign a letter if it was already assign to a user [2h, time spent: 2.5h]
backend frontend 3. priority
We'll need to add an empty field to the select field when assign a letter, and remove the letter's assignee accordingly in the API. ![image](https://user-images.githubusercontent.com/6004147/174788760-7fcfbb69-def3-4fb5-9de1-24435e82341d.png)
1.0
Allow staff to unassign a letter if it was already assign to a user [2h, time spent: 2.5h] - We'll need to add an empty field to the select field when assign a letter, and remove the letter's assignee accordingly in the API. ![image](https://user-images.githubusercontent.com/6004147/174788760-7fcfbb69-def3-4fb5-9de1-24435e82341d.png)
non_test
allow staff to unassign a letter if it was already assign to a user we ll need to add an empty field to the select field when assign a letter and remove the letter s assignee accordingly in the api
0
50,488
6,098,272,146
IssuesEvent
2017-06-20 07:02:09
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
closed
Points value goal
Ready to Test QA
Create a goal which is achieved when the user has reached a certain level of points in a specified points bucket This should have an option for whether it is cumulative (all points earned regardless of debits) or current balance
1.0
Points value goal - Create a goal which is achieved when the user has reached a certain level of points in a specified points bucket This should have an option for whether it is cumulative (all points earned regardless of debits) or current balance
test
points value goal create a goal which is achieved when the user has reached a certain level of points in a specified points bucket this should have an option for whether it is cumulative all points earned regardless of debits or current balance
1
10,184
3,089,390,129
IssuesEvent
2015-08-25 21:15:06
ES-DOC/esdoc-docs
https://api.github.com/repos/ES-DOC/esdoc-docs
closed
Display standard properties in the Viewer
ESPS feature request Viewer wip-awaiting-testing
WHO: ESPS This is the list of standard properties that must be displayed in the Model Component view for the ESPS project: 1. couplingFramework 2. shortName (you already use this) 3. longName (you already use this) 4. description (you already use this) 5. license 6. responsibleParty (you already use this?) 7. releaseDate 8. previousVersion 9. fundingSource 10. citation (you already use this) 11. onlineResource 12. componentLanguage 13. grid (this would be better off being shown in it's own tab) 14. type (you already use this?) 15. timing 16. documentAuthor (you already use this? but I'm not sure how it's different from responsibleParty) 17. documentCreationDate The following items need to be displayed from CIM 1.10. Ticket #92 must be completed first. 1. codeAccess 2. version
1.0
Display standard properties in the Viewer - WHO: ESPS This is the list of standard properties that must be displayed in the Model Component view for the ESPS project: 1. couplingFramework 2. shortName (you already use this) 3. longName (you already use this) 4. description (you already use this) 5. license 6. responsibleParty (you already use this?) 7. releaseDate 8. previousVersion 9. fundingSource 10. citation (you already use this) 11. onlineResource 12. componentLanguage 13. grid (this would be better off being shown in it's own tab) 14. type (you already use this?) 15. timing 16. documentAuthor (you already use this? but I'm not sure how it's different from responsibleParty) 17. documentCreationDate The following items need to be displayed from CIM 1.10. Ticket #92 must be completed first. 1. codeAccess 2. version
test
display standard properties in the viewer who esps this is the list of standard properties that must be displayed in the model component view for the esps project couplingframework shortname you already use this longname you already use this description you already use this license responsibleparty you already use this releasedate previousversion fundingsource citation you already use this onlineresource componentlanguage grid this would be better off being shown in it s own tab type you already use this timing documentauthor you already use this but i m not sure how it s different from responsibleparty documentcreationdate the following items need to be displayed from cim ticket must be completed first codeaccess version
1
94,347
10,819,953,710
IssuesEvent
2019-11-08 15:25:22
sandialabs/Gulliver
https://api.github.com/repos/sandialabs/Gulliver
opened
[Doc] Create a style guideline
documentation
# Description We're lacking a style guideline both for the C# code and the Sphinx doc generation. # Impact It will keep the code clean and concise for all devs # Proposed solutions - provide a document with the desired guidelines - settle upon a .editorconfig for all file types - create appropriate resharper code formatter - define linter / formatter rules as appropriate - run linter as part of build step
1.0
[Doc] Create a style guideline - # Description We're lacking a style guideline both for the C# code and the Sphinx doc generation. # Impact It will keep the code clean and concise for all devs # Proposed solutions - provide a document with the desired guidelines - settle upon a .editorconfig for all file types - create appropriate resharper code formatter - define linter / formatter rules as appropriate - run linter as part of build step
non_test
create a style guideline description we re lacking a style guideline both for the c code and the sphinx doc generation impact it will keep the code clean and concise for all devs proposed solutions provide a document with the desired guidelines settle upon a editorconfig for all file types create appropriate resharper code formatter define linter formatter rules as appropriate run linter as part of build step
0
203,282
15,360,477,418
IssuesEvent
2021-03-01 16:58:40
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Auto logout is inconsistently routing the session to the login page after session expiry
component:ui priority:low state:needs_test type:bug
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - UI ##### SUMMARY The Auto Logout functionality is not occuring consistently. I’m regularly thrown back to the login page if I’ve left a session idle <for some time> and I attempt to interact with an element of the page. - I've seen this on the Org Edit, Job Results Detail, and Settings License pages. - The page was in focus - single user signed in - no popup to notify the user that the session is expiring ##### ENVIRONMENT * AWX version: 3.0.0 Firefox and chrome ##### STEPS TO REPRODUCE None available. ##### EXPECTED RESULTS Session timeout modal to appear and the session automatically logged out. ##### ACTUAL RESULTS Session re-routed to the login screen after interacting with an element. ##### ADDITIONAL INFORMATION ![wfviz](https://user-images.githubusercontent.com/12446869/55257590-bb900400-5236-11e9-853a-aa3de1d05cb1.gif) [Archive 19-03-29 15-27-33.har.zip](https://github.com/ansible/awx/files/3024156/Archive.19-03-29.15-27-33.har.zip)
1.0
Auto logout is inconsistently routing the session to the login page after session expiry - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - UI ##### SUMMARY The Auto Logout functionality is not occuring consistently. I’m regularly thrown back to the login page if I’ve left a session idle <for some time> and I attempt to interact with an element of the page. - I've seen this on the Org Edit, Job Results Detail, and Settings License pages. - The page was in focus - single user signed in - no popup to notify the user that the session is expiring ##### ENVIRONMENT * AWX version: 3.0.0 Firefox and chrome ##### STEPS TO REPRODUCE None available. ##### EXPECTED RESULTS Session timeout modal to appear and the session automatically logged out. ##### ACTUAL RESULTS Session re-routed to the login screen after interacting with an element. ##### ADDITIONAL INFORMATION ![wfviz](https://user-images.githubusercontent.com/12446869/55257590-bb900400-5236-11e9-853a-aa3de1d05cb1.gif) [Archive 19-03-29 15-27-33.har.zip](https://github.com/ansible/awx/files/3024156/Archive.19-03-29.15-27-33.har.zip)
test
auto logout is inconsistently routing the session to the login page after session expiry issue type bug report component name ui summary the auto logout functionality is not occuring consistently i’m regularly thrown back to the login page if i’ve left a session idle and i attempt to interact with an element of the page i ve seen this on the org edit job results detail and settings license pages the page was in focus single user signed in no popup to notify the user that the session is expiring environment awx version firefox and chrome steps to reproduce none available expected results session timeout modal to appear and the session automatically logged out actual results session re routed to the login screen after interacting with an element additional information
1
247,528
7,919,528,255
IssuesEvent
2018-07-04 17:21:21
huridocs/uwazi
https://api.github.com/repos/huridocs/uwazi
closed
Publish action button being displayed in non-public items
Priority: Medium Question UI
![screen shot 2017-07-13 at 18 37 10](https://user-images.githubusercontent.com/489831/28193562-fb7175f4-67fa-11e7-98c8-c9bc0a710a1d.png) - Is this the desired behaviour? - Do we need to tweak the design of the cards for non pubic docs and entities?
1.0
Publish action button being displayed in non-public items - ![screen shot 2017-07-13 at 18 37 10](https://user-images.githubusercontent.com/489831/28193562-fb7175f4-67fa-11e7-98c8-c9bc0a710a1d.png) - Is this the desired behaviour? - Do we need to tweak the design of the cards for non pubic docs and entities?
non_test
publish action button being displayed in non public items is this the desired behaviour do we need to tweak the design of the cards for non pubic docs and entities
0
14,996
18,676,778,517
IssuesEvent
2021-10-31 17:43:14
slynch8/10x
https://api.github.com/repos/slynch8/10x
closed
Preprocessor doesn't recognise "true"
bug Priority 3 trivial preprocessor
![foo](https://user-images.githubusercontent.com/9505900/139592810-68940c15-b3a4-44fd-8e4b-7bf6047b16ea.png) If I use this instead: `#define FOO 1` Then the correct branch is used.
1.0
Preprocessor doesn't recognise "true" - ![foo](https://user-images.githubusercontent.com/9505900/139592810-68940c15-b3a4-44fd-8e4b-7bf6047b16ea.png) If I use this instead: `#define FOO 1` Then the correct branch is used.
non_test
preprocessor doesn t recognise true if i use this instead define foo then the correct branch is used
0
11,983
3,244,959,191
IssuesEvent
2015-10-16 07:32:48
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
На бэке (wf-base), в сущности Flow_ServiceData доработать учет департамента, как часть составного ключа
active hi priority test _wf-base
1) Скопировать сущность SubjectOrganDepartment из wf-central в wf-base (с классами и csv) ВАЖНО: в wf-base нужно: 1.1) информацию об этой новой сущности(ее создании в базе) прописывать в /changes/base_2015.xml (при этом, чтоб не было конфликта в wf-central(т.к. там эта сущность уже и так создана) - нужно changeSet по созданию сущности проставить failOnError="false") 1.2) подгрузку данных делать в: /load-data-local.xml /data/SubjectOrganDepartment.csv (т.к. в централе своя таблица с данными) 1.3) загрузить такие данные в в /data/SubjectOrganDepartment.csv: nID;sName;sGroup_Activiti;nID_SubjectOrgan 1;Test Department (_test_dnepr_mvd-1_auto);Test Group Activity;2 2;МРЕО,Киев,Петропавлівська Борщагівка, вул.Кільцева 4 (dnepr_mreo_1);demomreo;2 3;МРЕО,Киев,вул. Туполєва, 19 (kiev_mreo1_1);demomreo;2 4;МРЕО,Киев,вул. Велика кільцева дорога. 22-А (kiev_mreo2_1);demomreo;2 5;МРЕО,Киев,вул. Братиславська, 52 (kiev_mreо3_1);demomreo;2 6;МРЕО,Киев,вул. Столичне шосе, 104 (kiev_mreo4_1);demomreo;2 7;МРЕО,Киев,вул. Павла Усенка, 8 (kiev_mreo5_1);demomreo;2 8;МРЕО,Киев,вул. Новокостянтинівська, 8 (kiev_mreo6_1);demomreo;2 9;МРЕО,Киев,пров. Балтійський, 20 (kiev_mreo7_1);demomreo;2 10;МРЕО,Днепр (dnepr_mreo_1);demomreo;2 11;АВИР, Днепр, Поля 1 (dnepr_dms_passport);demoavir;2 2) Во всех сервисах работы с сущностью Flow_ServiceData и FlowProperty - добавить вспомогательный(опциональный) параметр nID_SubjectOrganDepartment, передавай который в паре с существующим sID_BP - будет выбираться не первая встретившаяся строка из сущности FlowProperty а именно та, в которой поле nID_SubjectOrganDepartment соответствует этому значению. Такие сервисы, например: /getFlowSlots_ServiceData /setFlowSlot_ServiceData /buildFlowSlots /clearFlowSlots /getSheduleFlowIncludes /getSheduleFlowExcludes /setSheduleFlowInclude /setSheduleFlowExclude /removeSheduleFlowInclude /removeSheduleFlowExclude (может еще что-то пропустил) 3) Добавить ключ в сущности на уникальность в сущности "Flow_ServiceData" пары "nID_SubjectOrganDepartment" и "sID_BP" 4) Обновить АПИ на нашей Вики
1.0
На бэке (wf-base), в сущности Flow_ServiceData доработать учет департамента, как часть составного ключа - 1) Скопировать сущность SubjectOrganDepartment из wf-central в wf-base (с классами и csv) ВАЖНО: в wf-base нужно: 1.1) информацию об этой новой сущности(ее создании в базе) прописывать в /changes/base_2015.xml (при этом, чтоб не было конфликта в wf-central(т.к. там эта сущность уже и так создана) - нужно changeSet по созданию сущности проставить failOnError="false") 1.2) подгрузку данных делать в: /load-data-local.xml /data/SubjectOrganDepartment.csv (т.к. в централе своя таблица с данными) 1.3) загрузить такие данные в в /data/SubjectOrganDepartment.csv: nID;sName;sGroup_Activiti;nID_SubjectOrgan 1;Test Department (_test_dnepr_mvd-1_auto);Test Group Activity;2 2;МРЕО,Киев,Петропавлівська Борщагівка, вул.Кільцева 4 (dnepr_mreo_1);demomreo;2 3;МРЕО,Киев,вул. Туполєва, 19 (kiev_mreo1_1);demomreo;2 4;МРЕО,Киев,вул. Велика кільцева дорога. 22-А (kiev_mreo2_1);demomreo;2 5;МРЕО,Киев,вул. Братиславська, 52 (kiev_mreо3_1);demomreo;2 6;МРЕО,Киев,вул. Столичне шосе, 104 (kiev_mreo4_1);demomreo;2 7;МРЕО,Киев,вул. Павла Усенка, 8 (kiev_mreo5_1);demomreo;2 8;МРЕО,Киев,вул. Новокостянтинівська, 8 (kiev_mreo6_1);demomreo;2 9;МРЕО,Киев,пров. Балтійський, 20 (kiev_mreo7_1);demomreo;2 10;МРЕО,Днепр (dnepr_mreo_1);demomreo;2 11;АВИР, Днепр, Поля 1 (dnepr_dms_passport);demoavir;2 2) Во всех сервисах работы с сущностью Flow_ServiceData и FlowProperty - добавить вспомогательный(опциональный) параметр nID_SubjectOrganDepartment, передавай который в паре с существующим sID_BP - будет выбираться не первая встретившаяся строка из сущности FlowProperty а именно та, в которой поле nID_SubjectOrganDepartment соответствует этому значению. Такие сервисы, например: /getFlowSlots_ServiceData /setFlowSlot_ServiceData /buildFlowSlots /clearFlowSlots /getSheduleFlowIncludes /getSheduleFlowExcludes /setSheduleFlowInclude /setSheduleFlowExclude /removeSheduleFlowInclude /removeSheduleFlowExclude (может еще что-то пропустил) 3) Добавить ключ в сущности на уникальность в сущности "Flow_ServiceData" пары "nID_SubjectOrganDepartment" и "sID_BP" 4) Обновить АПИ на нашей Вики
test
на бэке wf base в сущности flow servicedata доработать учет департамента как часть составного ключа скопировать сущность subjectorgandepartment из wf central в wf base с классами и csv важно в wf base нужно информацию об этой новой сущности ее создании в базе прописывать в changes base xml при этом чтоб не было конфликта в wf central т к там эта сущность уже и так создана нужно changeset по созданию сущности проставить failonerror false подгрузку данных делать в load data local xml data subjectorgandepartment csv т к в централе своя таблица с данными загрузить такие данные в в data subjectorgandepartment csv nid sname sgroup activiti nid subjectorgan test department test dnepr mvd auto test group activity мрео киев петропавлівська борщагівка вул кільцева dnepr mreo demomreo мрео киев вул туполєва kiev demomreo мрео киев вул велика кільцева дорога а kiev demomreo мрео киев вул братиславська kiev demomreo мрео киев вул столичне шосе kiev demomreo мрео киев вул павла усенка kiev demomreo мрео киев вул новокостянтинівська kiev demomreo мрео киев пров балтійський kiev demomreo мрео днепр dnepr mreo demomreo авир днепр поля dnepr dms passport demoavir во всех сервисах работы с сущностью flow servicedata и flowproperty добавить вспомогательный опциональный параметр nid subjectorgandepartment передавай который в паре с существующим sid bp будет выбираться не первая встретившаяся строка из сущности flowproperty а именно та в которой поле nid subjectorgandepartment соответствует этому значению такие сервисы например getflowslots servicedata setflowslot servicedata buildflowslots clearflowslots getsheduleflowincludes getsheduleflowexcludes setsheduleflowinclude setsheduleflowexclude removesheduleflowinclude removesheduleflowexclude может еще что то пропустил добавить ключ в сущности на уникальность в сущности flow servicedata пары nid subjectorgandepartment и sid bp обновить апи на нашей вики
1
82,794
16,040,893,032
IssuesEvent
2021-04-22 07:44:39
smeas/Beer-and-Plunder
https://api.github.com/repos/smeas/Beer-and-Plunder
closed
Set up basic brawl system
4p code
**Description** Set up the basic viking brawl system. **Subtasks** - [x] Vikings enter brawl state when reaching below a threshold - [x] Vikings can leave the brawl state when a condition is met - [x] There is an indicator to when a viking is brawling - [x] The brawling spreads to nearby tables - [ ] Entering a brawl causes damage to the tavern
1.0
Set up basic brawl system - **Description** Set up the basic viking brawl system. **Subtasks** - [x] Vikings enter brawl state when reaching below a threshold - [x] Vikings can leave the brawl state when a condition is met - [x] There is an indicator to when a viking is brawling - [x] The brawling spreads to nearby tables - [ ] Entering a brawl causes damage to the tavern
non_test
set up basic brawl system description set up the basic viking brawl system subtasks vikings enter brawl state when reaching below a threshold vikings can leave the brawl state when a condition is met there is an indicator to when a viking is brawling the brawling spreads to nearby tables entering a brawl causes damage to the tavern
0
90,004
3,808,110,819
IssuesEvent
2016-03-25 13:19:55
marklogic/marklogic-data-hub
https://api.github.com/repos/marklogic/marklogic-data-hub
closed
Add REST scaffolding
Component:DHIB Component:QuickStart priority:high
Need scaffolding to generate REST stuff. For REST extensions use the provided parameters from the sidecar xml to auto place let vars inside the code. Saves the end user some time.
1.0
Add REST scaffolding - Need scaffolding to generate REST stuff. For REST extensions use the provided parameters from the sidecar xml to auto place let vars inside the code. Saves the end user some time.
non_test
add rest scaffolding need scaffolding to generate rest stuff for rest extensions use the provided parameters from the sidecar xml to auto place let vars inside the code saves the end user some time
0
4,999
3,489,018,541
IssuesEvent
2016-01-03 14:19:32
SFTtech/openage
https://api.github.com/repos/SFTtech/openage
closed
make fails for every run after the first during generating the codegen depencies because of slp.cython.so
bug buildsystem
The following should reproduce it ./configure & make & make #or configure I get [...] Call Stack (most recent call first): CMakeLists.txt:72 (doxygen_configure) codegeneration depends on non-.py module openage/convert/slp.cpython-34.so CMake Error at buildsystem/codegen.cmake:25 (message): failed to get target list from codegen invocation Call Stack (most recent call first): libopenage/CMakeLists.txt:55 (codegen_run) [...] After some debugging by TheJJ it appears that during the first make run, the slp.cython.[46].so gets generated by cython and that is then taken into consideration on finding the dependencies during the second run and it fails that condition. I also want to add, that make mrproper doesn't remove these files (and some others like the gen'ed .cpp and .h files and some __init__.py files). Maybe this assumption is wrong, but in my naive view I would expect mrproper to clean everything added by the build probcess. This would be a different issue than this one though, if it even is one. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/29473449-make-fails-for-every-run-after-the-first-during-generating-the-codegen-depencies-because-of-slp-cython-so?utm_campaign=plugin&utm_content=tracker%2F371436&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F371436&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
make fails for every run after the first during generating the codegen depencies because of slp.cython.so - The following should reproduce it ./configure & make & make #or configure I get [...] Call Stack (most recent call first): CMakeLists.txt:72 (doxygen_configure) codegeneration depends on non-.py module openage/convert/slp.cpython-34.so CMake Error at buildsystem/codegen.cmake:25 (message): failed to get target list from codegen invocation Call Stack (most recent call first): libopenage/CMakeLists.txt:55 (codegen_run) [...] After some debugging by TheJJ it appears that during the first make run, the slp.cython.[46].so gets generated by cython and that is then taken into consideration on finding the dependencies during the second run and it fails that condition. I also want to add, that make mrproper doesn't remove these files (and some others like the gen'ed .cpp and .h files and some __init__.py files). Maybe this assumption is wrong, but in my naive view I would expect mrproper to clean everything added by the build probcess. This would be a different issue than this one though, if it even is one. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/29473449-make-fails-for-every-run-after-the-first-during-generating-the-codegen-depencies-because-of-slp-cython-so?utm_campaign=plugin&utm_content=tracker%2F371436&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F371436&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_test
make fails for every run after the first during generating the codegen depencies because of slp cython so the following should reproduce it configure make make or configure i get call stack most recent call first cmakelists txt doxygen configure codegeneration depends on non py module openage convert slp cpython so cmake error at buildsystem codegen cmake message failed to get target list from codegen invocation call stack most recent call first libopenage cmakelists txt codegen run after some debugging by thejj it appears that during the first make run the slp cython so gets generated by cython and that is then taken into consideration on finding the dependencies during the second run and it fails that condition i also want to add that make mrproper doesn t remove these files and some others like the gen ed cpp and h files and some init py files maybe this assumption is wrong but in my naive view i would expect mrproper to clean everything added by the build probcess this would be a different issue than this one though if it even is one want to back this issue we accept bounties via
0
285,148
24,645,945,861
IssuesEvent
2022-10-17 14:52:03
Kimthien94/Feedback-Online
https://api.github.com/repos/Kimthien94/Feedback-Online
opened
[BugID_26]_GUI_Tạo mới học viên_[Account học viên] textbox_Không cho phép nhập ký tự đặc biệt
bug comestic low Intergration test open UI_Position/Size
Thực tế: Không cho phép nhập ký tự đặc biệt Mong đợi: Cho phép nhập ký tự đặc biệt (Kiểu DL: String) Test case ID = 6
1.0
[BugID_26]_GUI_Tạo mới học viên_[Account học viên] textbox_Không cho phép nhập ký tự đặc biệt - Thực tế: Không cho phép nhập ký tự đặc biệt Mong đợi: Cho phép nhập ký tự đặc biệt (Kiểu DL: String) Test case ID = 6
test
gui tạo mới học viên textbox không cho phép nhập ký tự đặc biệt thực tế không cho phép nhập ký tự đặc biệt mong đợi cho phép nhập ký tự đặc biệt kiểu dl string test case id
1
2,125
2,586,365,769
IssuesEvent
2015-02-17 10:58:23
jplusplus/detective.io
https://api.github.com/repos/jplusplus/detective.io
closed
my user can't access to dio anymore
blocker bug need testing
If i try to access to Detective.io with my super user "annelise", i have a 500 error ![capture d ecran 2015-02-16 a 15 56 39](https://cloud.githubusercontent.com/assets/2103603/6213883/7dcfdcae-b5f4-11e4-86cd-df0d730958fa.png) ![capture d ecran 2015-02-16 a 15 56 10](https://cloud.githubusercontent.com/assets/2103603/6213884/7dd28e36-b5f4-11e4-913c-a758a85caa99.png) Everything works well with another accouny
1.0
my user can't access to dio anymore - If i try to access to Detective.io with my super user "annelise", i have a 500 error ![capture d ecran 2015-02-16 a 15 56 39](https://cloud.githubusercontent.com/assets/2103603/6213883/7dcfdcae-b5f4-11e4-86cd-df0d730958fa.png) ![capture d ecran 2015-02-16 a 15 56 10](https://cloud.githubusercontent.com/assets/2103603/6213884/7dd28e36-b5f4-11e4-913c-a758a85caa99.png) Everything works well with another accouny
test
my user can t access to dio anymore if i try to access to detective io with my super user annelise i have a error everything works well with another accouny
1
179,885
13,909,156,304
IssuesEvent
2020-10-20 14:36:43
forem/forem
https://api.github.com/repos/forem/forem
closed
Attempting to run RSpec on Docker triggers Elastic error
area: containers area: search/indexing area: tests type: bug
<!-- Before creating a bug report, try disabling browser extensions to see if the bug is still present. --> <!-- If you're having trouble updating your profile, it is likely because you logged in separately with GitHub & Twitter. Please check if this is the case before creating a bug report, and email yo@dev.to so we can merge your accounts. --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When I attempt to run a particular spec in Docker, there's a complaint about a missing VCR stub for elasticsearch. The spec I'm running: `docker exec -it forem_rails bundle exec rspec spec/liquid_tags/vimeo_tag_spec.rb` The error I get: ``` Tasks: TOP => db:test:load => db:test:purge => db:check_protected_environments (See full trace by running task with --trace) [Zonebie] Setting timezone: ZONEBIE_TZ="Prague" An error occurred in a `before(:suite)` hook. Failure/Error: next unless Search::Client.indices.exists(index: search_class::INDEX_NAME) VCR::Errors::UnhandledHTTPRequestError: ================================================================================ An HTTP request has been made that VCR does not know how to handle: HEAD http://elasticsearch:9200/chat_channel_memberships_test There is currently no cassette in use. There are a few ways you can configure VCR to handle this request: * If you're surprised VCR is raising this error and want insight about how VCR attempted to handle the request, you can use the debug_logger configuration option to log more details [1]. ``` **To Reproduce** 1. Install Docker on WSL2 (Ubuntu) (Not sure if this is present in other environments, test and see?) 1. Run `bin/container-setup` 1. While docker is up, run `docker exec -it forem_rails bundle exec rspec spec/liquid_tags/vimeo_tag_spec.rb` **Expected behavior** All test cases pass, and RSpec reports the output. **Screenshots** ![image](https://user-images.githubusercontent.com/8124558/95024884-d6291180-0653-11eb-83df-5d1fd5df1561.png) **Desktop (please complete the following information):** - Windows 10 > (WSL2) Ubuntu > Docker
1.0
Attempting to run RSpec on Docker triggers Elastic error - <!-- Before creating a bug report, try disabling browser extensions to see if the bug is still present. --> <!-- If you're having trouble updating your profile, it is likely because you logged in separately with GitHub & Twitter. Please check if this is the case before creating a bug report, and email yo@dev.to so we can merge your accounts. --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When I attempt to run a particular spec in Docker, there's a complaint about a missing VCR stub for elasticsearch. The spec I'm running: `docker exec -it forem_rails bundle exec rspec spec/liquid_tags/vimeo_tag_spec.rb` The error I get: ``` Tasks: TOP => db:test:load => db:test:purge => db:check_protected_environments (See full trace by running task with --trace) [Zonebie] Setting timezone: ZONEBIE_TZ="Prague" An error occurred in a `before(:suite)` hook. Failure/Error: next unless Search::Client.indices.exists(index: search_class::INDEX_NAME) VCR::Errors::UnhandledHTTPRequestError: ================================================================================ An HTTP request has been made that VCR does not know how to handle: HEAD http://elasticsearch:9200/chat_channel_memberships_test There is currently no cassette in use. There are a few ways you can configure VCR to handle this request: * If you're surprised VCR is raising this error and want insight about how VCR attempted to handle the request, you can use the debug_logger configuration option to log more details [1]. ``` **To Reproduce** 1. Install Docker on WSL2 (Ubuntu) (Not sure if this is present in other environments, test and see?) 1. Run `bin/container-setup` 1. While docker is up, run `docker exec -it forem_rails bundle exec rspec spec/liquid_tags/vimeo_tag_spec.rb` **Expected behavior** All test cases pass, and RSpec reports the output. **Screenshots** ![image](https://user-images.githubusercontent.com/8124558/95024884-d6291180-0653-11eb-83df-5d1fd5df1561.png) **Desktop (please complete the following information):** - Windows 10 > (WSL2) Ubuntu > Docker
test
attempting to run rspec on docker triggers elastic error describe the bug when i attempt to run a particular spec in docker there s a complaint about a missing vcr stub for elasticsearch the spec i m running docker exec it forem rails bundle exec rspec spec liquid tags vimeo tag spec rb the error i get tasks top db test load db test purge db check protected environments see full trace by running task with trace setting timezone zonebie tz prague an error occurred in a before suite hook failure error next unless search client indices exists index search class index name vcr errors unhandledhttprequesterror an http request has been made that vcr does not know how to handle head there is currently no cassette in use there are a few ways you can configure vcr to handle this request if you re surprised vcr is raising this error and want insight about how vcr attempted to handle the request you can use the debug logger configuration option to log more details to reproduce install docker on ubuntu not sure if this is present in other environments test and see run bin container setup while docker is up run docker exec it forem rails bundle exec rspec spec liquid tags vimeo tag spec rb expected behavior all test cases pass and rspec reports the output screenshots desktop please complete the following information windows ubuntu docker
1
160,962
12,527,867,120
IssuesEvent
2020-06-04 08:38:53
MPDL/KEEPER
https://api.github.com/repos/MPDL/KEEPER
closed
Low: Tombstone Page for Archiving
gui test
Description should be shortened (like the landing page). Currently the full text ist being shown.
1.0
Low: Tombstone Page for Archiving - Description should be shortened (like the landing page). Currently the full text ist being shown.
test
low tombstone page for archiving description should be shortened like the landing page currently the full text ist being shown
1
433,175
12,503,225,672
IssuesEvent
2020-06-02 06:47:18
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
opened
portal preview: issue when testing OAuth or OpenID sources
Priority: Medium Type: Bug
**Describe the bug** When you try to preview a portal with an OAuth or an OpenID source from web admin, redirection are wrong **To Reproduce** Steps to reproduce the behavior: 1. Create following OpenID source: ```ini [google] create_local_account=no client_secret=fakeapikey password_length=8 access_token_path=https://accounts.google.com/token authorize_path=https://accounts.google.com/o/oauth2/v2/auth set_access_durations_action= protected_resource_url=https://accounts.google.com scope=openid local_account_logins=0 client_id=packetfence description=google openid domains=https://accounts.google.com hash_passwords=bcrypt site=https://accounts.google.com redirect_url=https://pfcen7dev.lan/oauth2/callback type=OpenID ``` 2. Assign OpenID source to default connection profile 3. Preview After you accept AUP, portal displays: ``` Not Implemented GET to /o/oauth2/v2/auth not supported. ``` **Expected behavior** You should be redirected to https://accounts.google.com/o/oauth2/v2/auth, not to `/o/oauth2/v2/auth` **Additional context** Issue is not present if tests are done directly from captive portal.
1.0
portal preview: issue when testing OAuth or OpenID sources - **Describe the bug** When you try to preview a portal with an OAuth or an OpenID source from web admin, redirection are wrong **To Reproduce** Steps to reproduce the behavior: 1. Create following OpenID source: ```ini [google] create_local_account=no client_secret=fakeapikey password_length=8 access_token_path=https://accounts.google.com/token authorize_path=https://accounts.google.com/o/oauth2/v2/auth set_access_durations_action= protected_resource_url=https://accounts.google.com scope=openid local_account_logins=0 client_id=packetfence description=google openid domains=https://accounts.google.com hash_passwords=bcrypt site=https://accounts.google.com redirect_url=https://pfcen7dev.lan/oauth2/callback type=OpenID ``` 2. Assign OpenID source to default connection profile 3. Preview After you accept AUP, portal displays: ``` Not Implemented GET to /o/oauth2/v2/auth not supported. ``` **Expected behavior** You should be redirected to https://accounts.google.com/o/oauth2/v2/auth, not to `/o/oauth2/v2/auth` **Additional context** Issue is not present if tests are done directly from captive portal.
non_test
portal preview issue when testing oauth or openid sources describe the bug when you try to preview a portal with an oauth or an openid source from web admin redirection are wrong to reproduce steps to reproduce the behavior create following openid source ini create local account no client secret fakeapikey password length access token path authorize path set access durations action protected resource url scope openid local account logins client id packetfence description google openid domains hash passwords bcrypt site redirect url type openid assign openid source to default connection profile preview after you accept aup portal displays not implemented get to o auth not supported expected behavior you should be redirected to not to o auth additional context issue is not present if tests are done directly from captive portal
0
227,650
18,076,352,926
IssuesEvent
2021-09-21 10:18:55
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_fieldsMapping and com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_allTypes fails with "java.util.NoSuchElementException: No value present"
Team: Core Type: Test-Failure Source: Internal
com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_fieldsMapping and com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_allTypes fails with "java.util.NoSuchElementException: No value present" Master / 5.1 (462045c) Master Sonar http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-master-sonar/985/testReport/ Stacktrace: ``` java.util.NoSuchElementException: No value present at java.base/java.util.Optional.get(Optional.java:148) at com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.randomEntryFrom(SqlCompactTest.java:498) at com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_allTypes(SqlCompactTest.java:357) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:834) ``` Standard output: ``` Finished Running Test: when_explicitTopLevelField_then_fail_this in 0.003 seconds. Started Running Test: test_allTypes 22:40:54,350 DEBUG || - [CalciteSqlOptimizer] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Before logical opt: LogicalTableSink(table=[[hazelcast, public, o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]]]], operation=[INSERT], flattened=[false]) LogicalProject(id=[13], __key=[null:OBJECT], character=[_UTF-16LE'a':VARCHAR CHARACTER SET "UTF-16LE"], string=[$0], boolean=[$1], byte=[$2], short=[$3], int=[$4], long=[$5], float=[$6], double=[$7], decimal=[$8], time=[$9], date=[$10], timestamp=[$11], timestampTz=[$12], this=[null:OBJECT]) LogicalTableScan(table=[[hazelcast, public, o_92291d0c_e1ee_43a6_a76e_f05b59acbecf[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]]]]) 22:40:54,351 DEBUG || - [CalciteSqlOptimizer] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] After logical opt: SinkLogicalRel(table=[[hazelcast, public, o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]]]], operation=[INSERT], flattened=[false]) ProjectLogicalRel(id=[13], __key=[null:OBJECT], character=[_UTF-16LE'a':VARCHAR CHARACTER SET "UTF-16LE"], string=[$0], boolean=[$1], byte=[$2], short=[$3], int=[$4], long=[$5], float=[$6], double=[$7], decimal=[$8], time=[$9], date=[$10], timestamp=[$11], timestampTz=[$12], this=[null:OBJECT]) FullScanLogicalRel(table=[[hazelcast, public, o_92291d0c_e1ee_43a6_a76e_f05b59acbecf[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]]]]) 22:40:54,352 DEBUG || - [CalciteSqlOptimizer] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] After physical opt: SinkPhysicalRel(table=[[hazelcast, public, o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]]]], operation=[INSERT], flattened=[false]) ProjectPhysicalRel(id=[13], __key=[null:OBJECT], character=[_UTF-16LE'a':VARCHAR CHARACTER SET "UTF-16LE"], string=[$0], boolean=[$1], byte=[$2], short=[$3], int=[$4], long=[$5], float=[$6], double=[$7], decimal=[$8], time=[$9], date=[$10], timestamp=[$11], timestampTz=[$12], this=[null:OBJECT]) FullScanPhysicalRel(table=[[hazelcast, public, o_92291d0c_e1ee_43a6_a76e_f05b59acbecf[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]]]]) 22:40:54,354 DEBUG || - [LightMasterContext] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Start executing light job 06d1-b9b8-f284-0001, execution graph in DOT format: digraph DAG { "AllTypes[public.o_92291d0c_e1ee_43a6_a76e_f05b59acbecf]" [localParallelism=1]; "Project" [localParallelism=default]; "Project(IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9])" [localParallelism=default]; "IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]" [localParallelism=1]; "AllTypes[public.o_92291d0c_e1ee_43a6_a76e_f05b59acbecf]" -> "Project" [queueSize=1024]; "Project" -> "Project(IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9])" [queueSize=1024]; "Project(IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9])" -> "IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]" [queueSize=1024]; } HINT: You can use graphviz or http://viz-js.com to visualize the printed graph. 22:40:54,354 DEBUG || - [LightMasterContext] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Building execution plan for 06d1-b9b8-f284-0001 22:40:54,354 DEBUG || - [LightMasterContext] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Built execution plans for 06d1-b9b8-f284-0001 22:40:54,355 DEBUG || - [InitExecutionOperation] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Initializing execution plan for job 06d1-b9b8-f284-0001, execution 06d1-b9b8-f284-0001 from [127.0.0.1]:5701 22:40:54,358 DEBUG || - [JobExecutionService] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Execution plan for light job ID=06d1-b9b8-f284-0001, jobName='06d1-b9b8-f284-0001', executionId=06d1-b9b8-f284-0001 initialized, will start the execution 22:40:54,361 WARN || - [TaskletExecutionService] hz.naughty_dijkstra.jet.cooperative.thread-0 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Exception in ProcessorTasklet{06d1-b9b8-f284-0001/IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]#0} com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at com.hazelcast.internal.serialization.impl.SerializationUtil.handleSerializeException(SerializationUtil.java:129) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:238) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:214) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toData(AbstractSerializationService.java:199) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.toDataWithStrategy(MapProxySupport.java:1320) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.putAllInternal(MapProxySupport.java:997) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:498) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:112) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.submitPending(WriteMapP.java:139) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.processInternal(WriteMapP.java:124) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.AsyncHazelcastWriterP.process(AsyncHazelcastWriterP.java:75) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.lambda$processInbox$2f647568$2(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.function.RunnableEx.run(RunnableEx.java:31) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.util.Util.doWithClassLoader(Util.java:516) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.processInbox(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:305) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:300) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:281) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.call(ProcessorTasklet.java:255) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:388) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:803) [?:?] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:356) [hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.lang.Thread.run(Thread.java:834) [?:?] Caused by: java.lang.UnsupportedOperationException: The BETA compact format can only be used with 5.0 cluster at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.putAsync(MemberSchemaService.java:136) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.put(MemberSchemaService.java:131) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.putToSchemaService(CompactStreamSerializer.java:155) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.writeGenericRecord(CompactStreamSerializer.java:137) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:128) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:123) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.write(StreamSerializerAdapter.java:39) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:235) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 21 more 22:40:54,362 WARN || - [JobClassLoaderService] ForkJoinPool.commonPool-worker-5 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] JobClassLoaders for jobId=06d1-b9b8-f284-0001 already removed 22:40:54,362 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-5 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Completed execution of job '06d1-b9b8-f284-0001', execution 06d1-b9b8-f284-0001 22:40:54,362 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-5 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Execution of job '06d1-b9b8-f284-0001', execution 06d1-b9b8-f284-0001 completed with failure java.util.concurrent.CompletionException: com.hazelcast.jet.JetException: Exception in ProcessorTasklet{06d1-b9b8-f284-0001/IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]#0}: com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?] at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?] at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?] at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?] at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?] at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?] at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:489) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:420) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:406) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:803) ~[?:?] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:356) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.lang.Thread.run(Thread.java:834) ~[?:?] Caused by: com.hazelcast.jet.JetException: Exception in ProcessorTasklet{06d1-b9b8-f284-0001/IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]#0}: com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:400) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 3 more Caused by: com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at com.hazelcast.internal.serialization.impl.SerializationUtil.handleSerializeException(SerializationUtil.java:129) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:238) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:214) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toData(AbstractSerializationService.java:199) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.toDataWithStrategy(MapProxySupport.java:1320) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.putAllInternal(MapProxySupport.java:997) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:498) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:112) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.submitPending(WriteMapP.java:139) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.processInternal(WriteMapP.java:124) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.AsyncHazelcastWriterP.process(AsyncHazelcastWriterP.java:75) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.lambda$processInbox$2f647568$2(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.function.RunnableEx.run(RunnableEx.java:31) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.util.Util.doWithClassLoader(Util.java:516) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.processInbox(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:305) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:300) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:281) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.call(ProcessorTasklet.java:255) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:388) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 3 more Caused by: java.lang.UnsupportedOperationException: The BETA compact format can only be used with 5.0 cluster at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.putAsync(MemberSchemaService.java:136) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.put(MemberSchemaService.java:131) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.putToSchemaService(CompactStreamSerializer.java:155) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.writeGenericRecord(CompactStreamSerializer.java:137) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:128) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:123) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.write(StreamSerializerAdapter.java:39) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:235) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:214) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toData(AbstractSerializationService.java:199) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.toDataWithStrategy(MapProxySupport.java:1320) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.putAllInternal(MapProxySupport.java:997) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:498) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:112) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.submitPending(WriteMapP.java:139) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.processInternal(WriteMapP.java:124) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.AsyncHazelcastWriterP.process(AsyncHazelcastWriterP.java:75) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.lambda$processInbox$2f647568$2(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.function.RunnableEx.run(RunnableEx.java:31) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.util.Util.doWithClassLoader(Util.java:516) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.processInbox(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:305) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:300) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:281) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.call(ProcessorTasklet.java:255) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:388) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 3 more 22:40:54,477 INFO |test_allTypes| - [SqlTestSupport] Time-limited test - Removing 1 cached plans in SqlTestSupport.@After 22:40:54,478 INFO |test_allTypes| - [SimpleTestInClusterSupport] Time-limited test - Ditching 0 jobs in SimpleTestInClusterSupport.@After: [] 22:40:54,478 INFO |test_allTypes| - [SimpleTestInClusterSupport] Time-limited test - Destroying 2 distributed objects in SimpleTestInClusterSupport.@After: [hz:impl:replicatedMapService/__sql.catalog, hz:impl:mapService/o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9] 22:40:54,478 ERROR || - [EventServiceImpl] hz.naughty_dijkstra.event-22 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] hz.naughty_dijkstra.event-22 caught an exception while processing:com.hazelcast.spi.impl.eventservice.impl.LocalEventDispatcher@1ef292f9 java.lang.UnsupportedOperationException: SQL catalog must never be cleared - MapEvent{entryEventType=CLEAR_ALL, member=Member [127.0.0.1]:5701 - 13ba42d0-3198-4384-8456-7a667126187f this, name='__sql.catalog', numberOfEntriesAffected=2} at com.hazelcast.jet.sql.impl.schema.MappingStorage$EntryListenerAdapter.mapCleared(MappingStorage.java:160) ~[classes/:?] at com.hazelcast.replicatedmap.impl.ReplicatedMapEventPublishingService.dispatchEvent(ReplicatedMapEventPublishingService.java:132) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.replicatedmap.impl.ReplicatedMapService.dispatchEvent(ReplicatedMapService.java:232) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.spi.impl.eventservice.impl.LocalEventDispatcher.run(LocalEventDispatcher.java:64) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.util.executor.StripedExecutor$Worker.process(StripedExecutor.java:245) [hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.util.executor.StripedExecutor$Worker.run(StripedExecutor.java:228) [hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] BuildInfo right after test_allTypes(com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest): BuildInfo{version='5.1-SNAPSHOT', build='20210917', buildNumber=20210917, revision=462045c, enterprise=false, serializationVersion=1} Hiccups measured while running test 'test_allTypes(com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest):' 22:40:50, accumulated pauses: 732 ms, max pause: 108 ms, pauses over 1000 ms: 0 No metrics recorded during the test ```
1.0
com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_fieldsMapping and com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_allTypes fails with "java.util.NoSuchElementException: No value present" - com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_fieldsMapping and com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_allTypes fails with "java.util.NoSuchElementException: No value present" Master / 5.1 (462045c) Master Sonar http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-master-sonar/985/testReport/ Stacktrace: ``` java.util.NoSuchElementException: No value present at java.base/java.util.Optional.get(Optional.java:148) at com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.randomEntryFrom(SqlCompactTest.java:498) at com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest.test_allTypes(SqlCompactTest.java:357) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:834) ``` Standard output: ``` Finished Running Test: when_explicitTopLevelField_then_fail_this in 0.003 seconds. Started Running Test: test_allTypes 22:40:54,350 DEBUG || - [CalciteSqlOptimizer] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Before logical opt: LogicalTableSink(table=[[hazelcast, public, o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]]]], operation=[INSERT], flattened=[false]) LogicalProject(id=[13], __key=[null:OBJECT], character=[_UTF-16LE'a':VARCHAR CHARACTER SET "UTF-16LE"], string=[$0], boolean=[$1], byte=[$2], short=[$3], int=[$4], long=[$5], float=[$6], double=[$7], decimal=[$8], time=[$9], date=[$10], timestamp=[$11], timestampTz=[$12], this=[null:OBJECT]) LogicalTableScan(table=[[hazelcast, public, o_92291d0c_e1ee_43a6_a76e_f05b59acbecf[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]]]]) 22:40:54,351 DEBUG || - [CalciteSqlOptimizer] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] After logical opt: SinkLogicalRel(table=[[hazelcast, public, o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]]]], operation=[INSERT], flattened=[false]) ProjectLogicalRel(id=[13], __key=[null:OBJECT], character=[_UTF-16LE'a':VARCHAR CHARACTER SET "UTF-16LE"], string=[$0], boolean=[$1], byte=[$2], short=[$3], int=[$4], long=[$5], float=[$6], double=[$7], decimal=[$8], time=[$9], date=[$10], timestamp=[$11], timestampTz=[$12], this=[null:OBJECT]) FullScanLogicalRel(table=[[hazelcast, public, o_92291d0c_e1ee_43a6_a76e_f05b59acbecf[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]]]]) 22:40:54,352 DEBUG || - [CalciteSqlOptimizer] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] After physical opt: SinkPhysicalRel(table=[[hazelcast, public, o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]]]], operation=[INSERT], flattened=[false]) ProjectPhysicalRel(id=[13], __key=[null:OBJECT], character=[_UTF-16LE'a':VARCHAR CHARACTER SET "UTF-16LE"], string=[$0], boolean=[$1], byte=[$2], short=[$3], int=[$4], long=[$5], float=[$6], double=[$7], decimal=[$8], time=[$9], date=[$10], timestamp=[$11], timestampTz=[$12], this=[null:OBJECT]) FullScanPhysicalRel(table=[[hazelcast, public, o_92291d0c_e1ee_43a6_a76e_f05b59acbecf[projects=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]]]]) 22:40:54,354 DEBUG || - [LightMasterContext] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Start executing light job 06d1-b9b8-f284-0001, execution graph in DOT format: digraph DAG { "AllTypes[public.o_92291d0c_e1ee_43a6_a76e_f05b59acbecf]" [localParallelism=1]; "Project" [localParallelism=default]; "Project(IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9])" [localParallelism=default]; "IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]" [localParallelism=1]; "AllTypes[public.o_92291d0c_e1ee_43a6_a76e_f05b59acbecf]" -> "Project" [queueSize=1024]; "Project" -> "Project(IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9])" [queueSize=1024]; "Project(IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9])" -> "IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]" [queueSize=1024]; } HINT: You can use graphviz or http://viz-js.com to visualize the printed graph. 22:40:54,354 DEBUG || - [LightMasterContext] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Building execution plan for 06d1-b9b8-f284-0001 22:40:54,354 DEBUG || - [LightMasterContext] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Built execution plans for 06d1-b9b8-f284-0001 22:40:54,355 DEBUG || - [InitExecutionOperation] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Initializing execution plan for job 06d1-b9b8-f284-0001, execution 06d1-b9b8-f284-0001 from [127.0.0.1]:5701 22:40:54,358 DEBUG || - [JobExecutionService] hz.naughty_dijkstra.client-query.thread-2 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Execution plan for light job ID=06d1-b9b8-f284-0001, jobName='06d1-b9b8-f284-0001', executionId=06d1-b9b8-f284-0001 initialized, will start the execution 22:40:54,361 WARN || - [TaskletExecutionService] hz.naughty_dijkstra.jet.cooperative.thread-0 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Exception in ProcessorTasklet{06d1-b9b8-f284-0001/IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]#0} com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at com.hazelcast.internal.serialization.impl.SerializationUtil.handleSerializeException(SerializationUtil.java:129) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:238) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:214) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toData(AbstractSerializationService.java:199) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.toDataWithStrategy(MapProxySupport.java:1320) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.putAllInternal(MapProxySupport.java:997) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:498) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:112) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.submitPending(WriteMapP.java:139) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.processInternal(WriteMapP.java:124) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.AsyncHazelcastWriterP.process(AsyncHazelcastWriterP.java:75) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.lambda$processInbox$2f647568$2(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.function.RunnableEx.run(RunnableEx.java:31) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.util.Util.doWithClassLoader(Util.java:516) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.processInbox(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:305) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:300) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:281) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.call(ProcessorTasklet.java:255) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:388) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:803) [?:?] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:356) [hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.lang.Thread.run(Thread.java:834) [?:?] Caused by: java.lang.UnsupportedOperationException: The BETA compact format can only be used with 5.0 cluster at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.putAsync(MemberSchemaService.java:136) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.put(MemberSchemaService.java:131) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.putToSchemaService(CompactStreamSerializer.java:155) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.writeGenericRecord(CompactStreamSerializer.java:137) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:128) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:123) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.write(StreamSerializerAdapter.java:39) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:235) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 21 more 22:40:54,362 WARN || - [JobClassLoaderService] ForkJoinPool.commonPool-worker-5 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] JobClassLoaders for jobId=06d1-b9b8-f284-0001 already removed 22:40:54,362 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-5 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Completed execution of job '06d1-b9b8-f284-0001', execution 06d1-b9b8-f284-0001 22:40:54,362 DEBUG || - [JobExecutionService] ForkJoinPool.commonPool-worker-5 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] Execution of job '06d1-b9b8-f284-0001', execution 06d1-b9b8-f284-0001 completed with failure java.util.concurrent.CompletionException: com.hazelcast.jet.JetException: Exception in ProcessorTasklet{06d1-b9b8-f284-0001/IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]#0}: com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) ~[?:?] at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) ~[?:?] at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:870) ~[?:?] at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?] at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?] at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) ~[?:?] at com.hazelcast.jet.impl.util.NonCompletableFuture.internalCompleteExceptionally(NonCompletableFuture.java:72) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$ExecutionTracker.taskletDone(TaskletExecutionService.java:489) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.dismissTasklet(TaskletExecutionService.java:420) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:406) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.util.concurrent.CopyOnWriteArrayList.forEach(CopyOnWriteArrayList.java:803) ~[?:?] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.run(TaskletExecutionService.java:356) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at java.lang.Thread.run(Thread.java:834) ~[?:?] Caused by: com.hazelcast.jet.JetException: Exception in ProcessorTasklet{06d1-b9b8-f284-0001/IMap[public.o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9]#0}: com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:400) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 3 more Caused by: com.hazelcast.nio.serialization.HazelcastSerializationException: Failed to serialize 'com.hazelcast.internal.serialization.impl.compact.DeserializedGenericRecord' at com.hazelcast.internal.serialization.impl.SerializationUtil.handleSerializeException(SerializationUtil.java:129) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:238) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:214) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toData(AbstractSerializationService.java:199) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.toDataWithStrategy(MapProxySupport.java:1320) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.putAllInternal(MapProxySupport.java:997) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:498) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:112) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.submitPending(WriteMapP.java:139) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.processInternal(WriteMapP.java:124) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.AsyncHazelcastWriterP.process(AsyncHazelcastWriterP.java:75) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.lambda$processInbox$2f647568$2(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.function.RunnableEx.run(RunnableEx.java:31) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.util.Util.doWithClassLoader(Util.java:516) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.processInbox(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:305) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:300) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:281) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.call(ProcessorTasklet.java:255) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:388) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 3 more Caused by: java.lang.UnsupportedOperationException: The BETA compact format can only be used with 5.0 cluster at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.putAsync(MemberSchemaService.java:136) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.schema.MemberSchemaService.put(MemberSchemaService.java:131) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.putToSchemaService(CompactStreamSerializer.java:155) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.writeGenericRecord(CompactStreamSerializer.java:137) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:128) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.compact.CompactStreamSerializer.write(CompactStreamSerializer.java:123) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.write(StreamSerializerAdapter.java:39) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:235) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(AbstractSerializationService.java:214) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toData(AbstractSerializationService.java:199) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.toDataWithStrategy(MapProxySupport.java:1320) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxySupport.putAllInternal(MapProxySupport.java:997) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:498) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.map.impl.proxy.MapProxyImpl.putAllAsync(MapProxyImpl.java:112) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.submitPending(WriteMapP.java:139) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.WriteMapP.processInternal(WriteMapP.java:124) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.connector.AsyncHazelcastWriterP.process(AsyncHazelcastWriterP.java:75) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.lambda$processInbox$2f647568$2(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.function.RunnableEx.run(RunnableEx.java:31) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.util.Util.doWithClassLoader(Util.java:516) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.processInbox(ProcessorTasklet.java:440) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:305) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:300) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.stateMachineStep(ProcessorTasklet.java:281) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.ProcessorTasklet.call(ProcessorTasklet.java:255) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.jet.impl.execution.TaskletExecutionService$CooperativeWorker.runTasklet(TaskletExecutionService.java:388) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] ... 3 more 22:40:54,477 INFO |test_allTypes| - [SqlTestSupport] Time-limited test - Removing 1 cached plans in SqlTestSupport.@After 22:40:54,478 INFO |test_allTypes| - [SimpleTestInClusterSupport] Time-limited test - Ditching 0 jobs in SimpleTestInClusterSupport.@After: [] 22:40:54,478 INFO |test_allTypes| - [SimpleTestInClusterSupport] Time-limited test - Destroying 2 distributed objects in SimpleTestInClusterSupport.@After: [hz:impl:replicatedMapService/__sql.catalog, hz:impl:mapService/o_873dc252_2d9a_4bd2_9a2e_ca08aa2915c9] 22:40:54,478 ERROR || - [EventServiceImpl] hz.naughty_dijkstra.event-22 - [127.0.0.1]:5701 [dev] [5.1-SNAPSHOT] hz.naughty_dijkstra.event-22 caught an exception while processing:com.hazelcast.spi.impl.eventservice.impl.LocalEventDispatcher@1ef292f9 java.lang.UnsupportedOperationException: SQL catalog must never be cleared - MapEvent{entryEventType=CLEAR_ALL, member=Member [127.0.0.1]:5701 - 13ba42d0-3198-4384-8456-7a667126187f this, name='__sql.catalog', numberOfEntriesAffected=2} at com.hazelcast.jet.sql.impl.schema.MappingStorage$EntryListenerAdapter.mapCleared(MappingStorage.java:160) ~[classes/:?] at com.hazelcast.replicatedmap.impl.ReplicatedMapEventPublishingService.dispatchEvent(ReplicatedMapEventPublishingService.java:132) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.replicatedmap.impl.ReplicatedMapService.dispatchEvent(ReplicatedMapService.java:232) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.spi.impl.eventservice.impl.LocalEventDispatcher.run(LocalEventDispatcher.java:64) ~[hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.util.executor.StripedExecutor$Worker.process(StripedExecutor.java:245) [hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] at com.hazelcast.internal.util.executor.StripedExecutor$Worker.run(StripedExecutor.java:228) [hazelcast-5.1-SNAPSHOT.jar:5.1-SNAPSHOT] BuildInfo right after test_allTypes(com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest): BuildInfo{version='5.1-SNAPSHOT', build='20210917', buildNumber=20210917, revision=462045c, enterprise=false, serializationVersion=1} Hiccups measured while running test 'test_allTypes(com.hazelcast.jet.sql.impl.connector.map.SqlCompactTest):' 22:40:50, accumulated pauses: 732 ms, max pause: 108 ms, pauses over 1000 ms: 0 No metrics recorded during the test ```
test
com hazelcast jet sql impl connector map sqlcompacttest test fieldsmapping and com hazelcast jet sql impl connector map sqlcompacttest test alltypes fails with java util nosuchelementexception no value present com hazelcast jet sql impl connector map sqlcompacttest test fieldsmapping and com hazelcast jet sql impl connector map sqlcompacttest test alltypes fails with java util nosuchelementexception no value present master master sonar stacktrace java util nosuchelementexception no value present at java base java util optional get optional java at com hazelcast jet sql impl connector map sqlcompacttest randomentryfrom sqlcompacttest java at com hazelcast jet sql impl connector map sqlcompacttest test alltypes sqlcompacttest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java base java util concurrent futuretask run futuretask java at java base java lang thread run thread java standard output finished running test when explicittoplevelfield then fail this in seconds started running test test alltypes debug hz naughty dijkstra client query thread before logical opt logicaltablesink table operation flattened logicalproject id key character string boolean byte short int long float double decimal time date timestamp timestamptz this logicaltablescan table debug hz naughty dijkstra client query thread after logical opt sinklogicalrel table operation flattened projectlogicalrel id key character string boolean byte short int long float double decimal time date timestamp timestamptz this fullscanlogicalrel table debug hz naughty dijkstra client query thread after physical opt sinkphysicalrel table operation flattened projectphysicalrel id key character string boolean byte short int long float double decimal time date timestamp timestamptz this fullscanphysicalrel table debug hz naughty dijkstra client query thread start executing light job execution graph in dot format digraph dag alltypes project project imap imap alltypes project project project imap project imap imap hint you can use graphviz or to visualize the printed graph debug hz naughty dijkstra client query thread building execution plan for debug hz naughty dijkstra client query thread built execution plans for debug hz naughty dijkstra client query thread initializing execution plan for job execution from debug hz naughty dijkstra client query thread execution plan for light job id jobname executionid initialized will start the execution warn hz naughty dijkstra jet cooperative thread exception in processortasklet imap com hazelcast nio serialization hazelcastserializationexception failed to serialize com hazelcast internal serialization impl compact deserializedgenericrecord at com hazelcast internal serialization impl serializationutil handleserializeexception serializationutil java at com hazelcast internal serialization impl abstractserializationservice tobytes abstractserializationservice java at com hazelcast internal serialization impl abstractserializationservice tobytes abstractserializationservice java at com hazelcast internal serialization impl abstractserializationservice todata abstractserializationservice java at com hazelcast map impl proxy mapproxysupport todatawithstrategy mapproxysupport java at com hazelcast map impl proxy mapproxysupport putallinternal mapproxysupport java at com hazelcast map impl proxy mapproxyimpl putallasync mapproxyimpl java at com hazelcast map impl proxy mapproxyimpl putallasync mapproxyimpl java at com hazelcast jet impl connector writemapp submitpending writemapp java at com hazelcast jet impl connector writemapp processinternal writemapp java at com hazelcast jet impl connector asynchazelcastwriterp process asynchazelcastwriterp java at com hazelcast jet impl execution processortasklet lambda processinbox processortasklet java at com hazelcast jet function runnableex run runnableex java at com hazelcast jet impl util util dowithclassloader util java at com hazelcast jet impl execution processortasklet processinbox processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet call processortasklet java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java at java util concurrent copyonwritearraylist foreach copyonwritearraylist java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker run taskletexecutionservice java at java lang thread run thread java caused by java lang unsupportedoperationexception the beta compact format can only be used with cluster at com hazelcast internal serialization impl compact schema memberschemaservice putasync memberschemaservice java at com hazelcast internal serialization impl compact schema memberschemaservice put memberschemaservice java at com hazelcast internal serialization impl compact compactstreamserializer puttoschemaservice compactstreamserializer java at com hazelcast internal serialization impl compact compactstreamserializer writegenericrecord compactstreamserializer java at com hazelcast internal serialization impl compact compactstreamserializer write compactstreamserializer java at com hazelcast internal serialization impl compact compactstreamserializer write compactstreamserializer java at com hazelcast internal serialization impl streamserializeradapter write streamserializeradapter java at com hazelcast internal serialization impl abstractserializationservice tobytes abstractserializationservice java more warn forkjoinpool commonpool worker jobclassloaders for jobid already removed debug forkjoinpool commonpool worker completed execution of job execution debug forkjoinpool commonpool worker execution of job execution completed with failure java util concurrent completionexception com hazelcast jet jetexception exception in processortasklet imap com hazelcast nio serialization hazelcastserializationexception failed to serialize com hazelcast internal serialization impl compact deserializedgenericrecord at java util concurrent completablefuture encodethrowable completablefuture java at java util concurrent completablefuture completethrowable completablefuture java at java util concurrent completablefuture uniwhencomplete completablefuture java at java util concurrent completablefuture uniwhencomplete tryfire completablefuture java at java util concurrent completablefuture postcomplete completablefuture java at java util concurrent completablefuture completeexceptionally completablefuture java at com hazelcast jet impl util noncompletablefuture internalcompleteexceptionally noncompletablefuture java at com hazelcast jet impl execution taskletexecutionservice executiontracker taskletdone taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker dismisstasklet taskletexecutionservice java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java at java util concurrent copyonwritearraylist foreach copyonwritearraylist java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker run taskletexecutionservice java at java lang thread run thread java caused by com hazelcast jet jetexception exception in processortasklet imap com hazelcast nio serialization hazelcastserializationexception failed to serialize com hazelcast internal serialization impl compact deserializedgenericrecord at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java more caused by com hazelcast nio serialization hazelcastserializationexception failed to serialize com hazelcast internal serialization impl compact deserializedgenericrecord at com hazelcast internal serialization impl serializationutil handleserializeexception serializationutil java at com hazelcast internal serialization impl abstractserializationservice tobytes abstractserializationservice java at com hazelcast internal serialization impl abstractserializationservice tobytes abstractserializationservice java at com hazelcast internal serialization impl abstractserializationservice todata abstractserializationservice java at com hazelcast map impl proxy mapproxysupport todatawithstrategy mapproxysupport java at com hazelcast map impl proxy mapproxysupport putallinternal mapproxysupport java at com hazelcast map impl proxy mapproxyimpl putallasync mapproxyimpl java at com hazelcast map impl proxy mapproxyimpl putallasync mapproxyimpl java at com hazelcast jet impl connector writemapp submitpending writemapp java at com hazelcast jet impl connector writemapp processinternal writemapp java at com hazelcast jet impl connector asynchazelcastwriterp process asynchazelcastwriterp java at com hazelcast jet impl execution processortasklet lambda processinbox processortasklet java at com hazelcast jet function runnableex run runnableex java at com hazelcast jet impl util util dowithclassloader util java at com hazelcast jet impl execution processortasklet processinbox processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet call processortasklet java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java more caused by java lang unsupportedoperationexception the beta compact format can only be used with cluster at com hazelcast internal serialization impl compact schema memberschemaservice putasync memberschemaservice java at com hazelcast internal serialization impl compact schema memberschemaservice put memberschemaservice java at com hazelcast internal serialization impl compact compactstreamserializer puttoschemaservice compactstreamserializer java at com hazelcast internal serialization impl compact compactstreamserializer writegenericrecord compactstreamserializer java at com hazelcast internal serialization impl compact compactstreamserializer write compactstreamserializer java at com hazelcast internal serialization impl compact compactstreamserializer write compactstreamserializer java at com hazelcast internal serialization impl streamserializeradapter write streamserializeradapter java at com hazelcast internal serialization impl abstractserializationservice tobytes abstractserializationservice java at com hazelcast internal serialization impl abstractserializationservice tobytes abstractserializationservice java at com hazelcast internal serialization impl abstractserializationservice todata abstractserializationservice java at com hazelcast map impl proxy mapproxysupport todatawithstrategy mapproxysupport java at com hazelcast map impl proxy mapproxysupport putallinternal mapproxysupport java at com hazelcast map impl proxy mapproxyimpl putallasync mapproxyimpl java at com hazelcast map impl proxy mapproxyimpl putallasync mapproxyimpl java at com hazelcast jet impl connector writemapp submitpending writemapp java at com hazelcast jet impl connector writemapp processinternal writemapp java at com hazelcast jet impl connector asynchazelcastwriterp process asynchazelcastwriterp java at com hazelcast jet impl execution processortasklet lambda processinbox processortasklet java at com hazelcast jet function runnableex run runnableex java at com hazelcast jet impl util util dowithclassloader util java at com hazelcast jet impl execution processortasklet processinbox processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet statemachinestep processortasklet java at com hazelcast jet impl execution processortasklet call processortasklet java at com hazelcast jet impl execution taskletexecutionservice cooperativeworker runtasklet taskletexecutionservice java more info test alltypes time limited test removing cached plans in sqltestsupport after info test alltypes time limited test ditching jobs in simpletestinclustersupport after info test alltypes time limited test destroying distributed objects in simpletestinclustersupport after error hz naughty dijkstra event hz naughty dijkstra event caught an exception while processing com hazelcast spi impl eventservice impl localeventdispatcher java lang unsupportedoperationexception sql catalog must never be cleared mapevent entryeventtype clear all member member this name sql catalog numberofentriesaffected at com hazelcast jet sql impl schema mappingstorage entrylisteneradapter mapcleared mappingstorage java at com hazelcast replicatedmap impl replicatedmapeventpublishingservice dispatchevent replicatedmapeventpublishingservice java at com hazelcast replicatedmap impl replicatedmapservice dispatchevent replicatedmapservice java at com hazelcast spi impl eventservice impl localeventdispatcher run localeventdispatcher java at com hazelcast internal util executor stripedexecutor worker process stripedexecutor java at com hazelcast internal util executor stripedexecutor worker run stripedexecutor java buildinfo right after test alltypes com hazelcast jet sql impl connector map sqlcompacttest buildinfo version snapshot build buildnumber revision enterprise false serializationversion hiccups measured while running test test alltypes com hazelcast jet sql impl connector map sqlcompacttest accumulated pauses ms max pause ms pauses over ms no metrics recorded during the test
1
159,299
20,048,347,408
IssuesEvent
2022-02-03 01:07:43
kapseliboi/crowdfunding-frontend
https://api.github.com/repos/kapseliboi/crowdfunding-frontend
opened
CVE-2018-3739 (High) detected in https-proxy-agent-1.0.0.tgz, https-proxy-agent-2.1.1.tgz
security vulnerability
## CVE-2018-3739 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>https-proxy-agent-1.0.0.tgz</b>, <b>https-proxy-agent-2.1.1.tgz</b></p></summary> <p> <details><summary><b>https-proxy-agent-1.0.0.tgz</b></p></summary> <p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p> <p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/https-proxy-agent/package.json</p> <p> Dependency Hierarchy: - wdio-sauce-service-0.3.1.tgz (Root Library) - sauce-connect-launcher-1.2.2.tgz - :x: **https-proxy-agent-1.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>https-proxy-agent-2.1.1.tgz</b></p></summary> <p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p> <p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.1.1.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.1.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/github/node_modules/https-proxy-agent/package.json</p> <p> Dependency Hierarchy: - next-4.2.3.tgz (Root Library) - glob-promise-3.2.0.tgz - semantic-release-8.2.3.tgz - github-12.1.0.tgz - :x: **https-proxy-agent-2.1.1.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> https-proxy-agent before 2.1.1 passes auth option to the Buffer constructor without proper sanitization, resulting in DoS and uninitialized memory leak in setups where an attacker could submit typed input to the 'auth' parameter (e.g. JSON). <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3739>CVE-2018-3739</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3739">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3739</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution (https-proxy-agent): 2.2.0</p> <p>Direct dependency fix Resolution (wdio-sauce-service): 0.4.2</p><p>Fix Resolution (https-proxy-agent): 2.2.0</p> <p>Direct dependency fix Resolution (next): 4.3.0-canary.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-3739 (High) detected in https-proxy-agent-1.0.0.tgz, https-proxy-agent-2.1.1.tgz - ## CVE-2018-3739 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>https-proxy-agent-1.0.0.tgz</b>, <b>https-proxy-agent-2.1.1.tgz</b></p></summary> <p> <details><summary><b>https-proxy-agent-1.0.0.tgz</b></p></summary> <p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p> <p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/https-proxy-agent/package.json</p> <p> Dependency Hierarchy: - wdio-sauce-service-0.3.1.tgz (Root Library) - sauce-connect-launcher-1.2.2.tgz - :x: **https-proxy-agent-1.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>https-proxy-agent-2.1.1.tgz</b></p></summary> <p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p> <p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.1.1.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.1.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/github/node_modules/https-proxy-agent/package.json</p> <p> Dependency Hierarchy: - next-4.2.3.tgz (Root Library) - glob-promise-3.2.0.tgz - semantic-release-8.2.3.tgz - github-12.1.0.tgz - :x: **https-proxy-agent-2.1.1.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> https-proxy-agent before 2.1.1 passes auth option to the Buffer constructor without proper sanitization, resulting in DoS and uninitialized memory leak in setups where an attacker could submit typed input to the 'auth' parameter (e.g. JSON). <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3739>CVE-2018-3739</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3739">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3739</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution (https-proxy-agent): 2.2.0</p> <p>Direct dependency fix Resolution (wdio-sauce-service): 0.4.2</p><p>Fix Resolution (https-proxy-agent): 2.2.0</p> <p>Direct dependency fix Resolution (next): 4.3.0-canary.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in https proxy agent tgz https proxy agent tgz cve high severity vulnerability vulnerable libraries https proxy agent tgz https proxy agent tgz https proxy agent tgz an http s proxy http agent implementation for https library home page a href path to dependency file package json path to vulnerable library node modules https proxy agent package json dependency hierarchy wdio sauce service tgz root library sauce connect launcher tgz x https proxy agent tgz vulnerable library https proxy agent tgz an http s proxy http agent implementation for https library home page a href path to dependency file package json path to vulnerable library node modules github node modules https proxy agent package json dependency hierarchy next tgz root library glob promise tgz semantic release tgz github tgz x https proxy agent tgz vulnerable library found in base branch master vulnerability details https proxy agent before passes auth option to the buffer constructor without proper sanitization resulting in dos and uninitialized memory leak in setups where an attacker could submit typed input to the auth parameter e g json publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution https proxy agent direct dependency fix resolution wdio sauce service fix resolution https proxy agent direct dependency fix resolution next canary step up your open source security game with whitesource
0
292,825
8,969,028,594
IssuesEvent
2019-01-29 09:45:35
HRZNStudio/Spatial
https://api.github.com/repos/HRZNStudio/Spatial
opened
Basic chunk decoration
priority: low type: feature type: worker
After getting primed, chunks should be decorated (trees, plants, ...) If structures is to much work for this, it should be split to a separate issue when investigating implementation. Using vanilla chunk loading safety mechanisms is fine (decorating at x and y + 8, to avoid cascading)
1.0
Basic chunk decoration - After getting primed, chunks should be decorated (trees, plants, ...) If structures is to much work for this, it should be split to a separate issue when investigating implementation. Using vanilla chunk loading safety mechanisms is fine (decorating at x and y + 8, to avoid cascading)
non_test
basic chunk decoration after getting primed chunks should be decorated trees plants if structures is to much work for this it should be split to a separate issue when investigating implementation using vanilla chunk loading safety mechanisms is fine decorating at x and y to avoid cascading
0
56,127
8,052,123,084
IssuesEvent
2018-08-01 18:16:45
EFForg/privacybadger
https://api.github.com/repos/EFForg/privacybadger
closed
Privacy Badger website indicates that Avast does not work with it
documentation & specs question website
I have both the Avast antivirus extension and PrivacyBadger installed, and have never had any issues with it. Are others still having issues? Per the last question on this link: https://www.eff.org/privacybadger
1.0
Privacy Badger website indicates that Avast does not work with it - I have both the Avast antivirus extension and PrivacyBadger installed, and have never had any issues with it. Are others still having issues? Per the last question on this link: https://www.eff.org/privacybadger
non_test
privacy badger website indicates that avast does not work with it i have both the avast antivirus extension and privacybadger installed and have never had any issues with it are others still having issues per the last question on this link
0
239,279
19,843,495,018
IssuesEvent
2022-01-21 01:37:27
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[Failing test] aks-engine-windows-containerd-master
sig/windows kind/failing-test triage/accepted
### Which jobs are failing? sig-release-master-informing, aks-engine-windows-containerd-master ### Which tests are failing? * ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd.Overall * kubetest.Up ### Since when has it been failing? 2022-01-18 18:55 CET ### Testgrid link https://testgrid.k8s.io/sig-release-master-informing#aks-engine-windows-containerd-master ### Reason for failure (if possible) [lastr run](https://prow.k8s.io/view/gs/kubernetes-jenkins/logs/ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd/1483770098157096960) ```bash e2e.go: Up expand_less | 1h0m17s -- | -- error creating cluster: cannot deploy: cannot get the create deployment future response: Future#WaitForCompletion: context has been cancelled: StatusCode=200 -- Original Error: context deadline exceeded ``` quick suspicion docker related, maybe because of docker shim removal. ### Anything else we need to know? _No response_ ### Relevant SIG(s) /sig windows cc @jsturtevant
1.0
[Failing test] aks-engine-windows-containerd-master - ### Which jobs are failing? sig-release-master-informing, aks-engine-windows-containerd-master ### Which tests are failing? * ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd.Overall * kubetest.Up ### Since when has it been failing? 2022-01-18 18:55 CET ### Testgrid link https://testgrid.k8s.io/sig-release-master-informing#aks-engine-windows-containerd-master ### Reason for failure (if possible) [lastr run](https://prow.k8s.io/view/gs/kubernetes-jenkins/logs/ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd/1483770098157096960) ```bash e2e.go: Up expand_less | 1h0m17s -- | -- error creating cluster: cannot deploy: cannot get the create deployment future response: Future#WaitForCompletion: context has been cancelled: StatusCode=200 -- Original Error: context deadline exceeded ``` quick suspicion docker related, maybe because of docker shim removal. ### Anything else we need to know? _No response_ ### Relevant SIG(s) /sig windows cc @jsturtevant
test
aks engine windows containerd master which jobs are failing sig release master informing aks engine windows containerd master which tests are failing ci kubernetes aks engine azure master windows containerd overall kubetest up since when has it been failing cet testgrid link reason for failure if possible bash go up expand less error creating cluster cannot deploy cannot get the create deployment future response future waitforcompletion context has been cancelled statuscode original error context deadline exceeded quick suspicion docker related maybe because of docker shim removal anything else we need to know no response relevant sig s sig windows cc jsturtevant
1
86,961
8,055,068,892
IssuesEvent
2018-08-02 08:07:47
pints-team/pints
https://api.github.com/repos/pints-team/pints
closed
Add utility function to "GROUP BY" in functional testing
feature functional-testing
This fragment: https://github.com/pints-team/functional-testing/blob/master/pfunk/plots/opt_fn_plot_by_commit.py#L51-L67 should be in its own utility function somewhere
1.0
Add utility function to "GROUP BY" in functional testing - This fragment: https://github.com/pints-team/functional-testing/blob/master/pfunk/plots/opt_fn_plot_by_commit.py#L51-L67 should be in its own utility function somewhere
test
add utility function to group by in functional testing this fragment should be in its own utility function somewhere
1
13,446
15,864,351,354
IssuesEvent
2021-04-08 13:44:05
Crocoblock/suggestions
https://api.github.com/repos/Crocoblock/suggestions
closed
JetEngine: compatibility with RankMath SEO Plugin
JetEngine compatibility
RankMath SEO Plugin doesn't see what I write on JetEngine's custom fields. Thus I opened a ticket about it on RankMath support and they said you need to add their API on your system. Could you please integrate it? Following link has more detail: https://support.rankmath.com/ticket/rank-math-wont-read-custom-fields-by-jet-engine/?view=all **from ticket# 9933
True
JetEngine: compatibility with RankMath SEO Plugin - RankMath SEO Plugin doesn't see what I write on JetEngine's custom fields. Thus I opened a ticket about it on RankMath support and they said you need to add their API on your system. Could you please integrate it? Following link has more detail: https://support.rankmath.com/ticket/rank-math-wont-read-custom-fields-by-jet-engine/?view=all **from ticket# 9933
non_test
jetengine compatibility with rankmath seo plugin rankmath seo plugin doesn t see what i write on jetengine s custom fields thus i opened a ticket about it on rankmath support and they said you need to add their api on your system could you please integrate it following link has more detail from ticket
0
96,356
8,607,377,479
IssuesEvent
2018-11-17 21:44:45
DinoDevs/GladiatusCrazyAddon
https://api.github.com/repos/DinoDevs/GladiatusCrazyAddon
closed
[Feature Request] Save gladiators to attack them on arena(s)
feature request testing
The idea behind this is to make a check of specific players to attack, when the opportunity appears on provinciarum. Can also be used to make a follow of those players that seems to be quite actives in Gladiatus. It can be made in any of the following two options: 1. The possibility to add players (nickname and province) on a small window, and save them. When you enter the arena or turma and, someone of the 5 players to attack matches any of the players added previously; make the name glimmer, or anything else to let the player know there is a match. 2. The possibility to make a 2nd "notes" windows, that pops when you enter turma or arena, where you can make notes. Any posteriors idea are welcome.
1.0
[Feature Request] Save gladiators to attack them on arena(s) - The idea behind this is to make a check of specific players to attack, when the opportunity appears on provinciarum. Can also be used to make a follow of those players that seems to be quite actives in Gladiatus. It can be made in any of the following two options: 1. The possibility to add players (nickname and province) on a small window, and save them. When you enter the arena or turma and, someone of the 5 players to attack matches any of the players added previously; make the name glimmer, or anything else to let the player know there is a match. 2. The possibility to make a 2nd "notes" windows, that pops when you enter turma or arena, where you can make notes. Any posteriors idea are welcome.
test
save gladiators to attack them on arena s the idea behind this is to make a check of specific players to attack when the opportunity appears on provinciarum can also be used to make a follow of those players that seems to be quite actives in gladiatus it can be made in any of the following two options the possibility to add players nickname and province on a small window and save them when you enter the arena or turma and someone of the players to attack matches any of the players added previously make the name glimmer or anything else to let the player know there is a match the possibility to make a notes windows that pops when you enter turma or arena where you can make notes any posteriors idea are welcome
1
81,304
7,778,261,498
IssuesEvent
2018-06-05 13:45:03
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
opened
SqlClient TestNotSupportedExceptionForTransactionScopeAsync test failing after async fix
area-System.Data.SqlClient test bug
The TestNotSupportedExceptionForTransactionScopeAsync test is expecting a NotSupportedException to be thrown. The test was buggy, and wasn't awaiting the Assert.ThrowsAsync. Now that it's been fixed to await it, the test is failing because it's expecting a NotSupportedException that's not being thrown. I'm not sure what the desired behavior is here (maybe the test should just be deleted as transaction scope is supported now?), so I've left the test disabled.
1.0
SqlClient TestNotSupportedExceptionForTransactionScopeAsync test failing after async fix - The TestNotSupportedExceptionForTransactionScopeAsync test is expecting a NotSupportedException to be thrown. The test was buggy, and wasn't awaiting the Assert.ThrowsAsync. Now that it's been fixed to await it, the test is failing because it's expecting a NotSupportedException that's not being thrown. I'm not sure what the desired behavior is here (maybe the test should just be deleted as transaction scope is supported now?), so I've left the test disabled.
test
sqlclient testnotsupportedexceptionfortransactionscopeasync test failing after async fix the testnotsupportedexceptionfortransactionscopeasync test is expecting a notsupportedexception to be thrown the test was buggy and wasn t awaiting the assert throwsasync now that it s been fixed to await it the test is failing because it s expecting a notsupportedexception that s not being thrown i m not sure what the desired behavior is here maybe the test should just be deleted as transaction scope is supported now so i ve left the test disabled
1
153,958
12,178,657,209
IssuesEvent
2020-04-28 09:21:55
elastic/cloud-on-k8s
https://api.github.com/repos/elastic/cloud-on-k8s
closed
Error 500 in E2E test TestTelemetry
>flaky_test
https://devops-ci.elastic.co/job/cloud-on-k8s-e2e-tests-snapshot-versions/42/testReport/github/com_elastic_cloud-on-k8s_test_e2e_kb/TestTelemetry_Kibana_should_expose_eck_info_in_telemetry_data/ ``` === RUN TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data: telemetry_test.go:46: Error Trace: telemetry_test.go:46 Error: Received unexpected error: fail to request /api/telemetry/v2/clusters/_stats, status is 500) Test: TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data {"log.level":"error","@timestamp":"2020-04-28T03:27:38.944Z","message":"stopping early","service.version":"0.0.0-00000000","service.type":"eck","ecs.version":"1.4.0","error":"test failure","error.stack_trace":"github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/pkg/mod/github.com/go-logr/zapr@v0.1.0/zapr.go:128\ngithub.com/elastic/cloud-on-k8s/test/e2e/test.StepList.RunSequential\n\t/go/src/github.com/elastic/cloud-on-k8s/test/e2e/test/step.go:43\ngithub.com/elastic/cloud-on-k8s/test/e2e/kb.TestTelemetry\n\t/go/src/github.com/elastic/cloud-on-k8s/test/e2e/kb/telemetry_test.go:60\ntesting.tRunner\n\t/usr/local/go/src/testing/testing.go:991"} --- FAIL: TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data (0.13s) ``` Kibana logs report a failure on the `.security_7` shard: ``` {"type":"error","@timestamp":"2020-04-28T03:27:38Z","tags":[],"pid":6,"level":"error","error":{"message":"[unavailable_shards_exception] at least one primary shard for the index [.security-7] is unavailable","name":"Error","stack":"Error: [unavailable_shards_exception] at least one primary shard for the index [.security-7] is unavailable\n at respond (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:349:15)\n at checkRespForFailure (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:306:7)\n at HttpConnector.<anonymous> (/usr/share/kibana/node_modules/elasticsearch/src/lib/connectors/http.js:173:7)\n at IncomingMessage.wrapper (/usr/share/kibana/node_modules/elasticsearch/node_modules/lodash/lodash.js:4929:19)\n at IncomingMessage.emit (events.js:203:15)\n at endReadableNT (_stream_readable.js:1145:12)\n at process._tickCallback (internal/process/next_tick.js:63:19)"},"url":{"protocol":null,"slashes":null,"auth":null,"host":null,"port":null,"hostname":null,"hash":null,"search":null,"query":{},"pathname":"/api/telemetry/v2/clusters/_stats","path":"/api/telemetry/v2/clusters/_stats","href":"/api/telemetry/v2/clusters/_stats"},"message":"[unavailable_shards_exception] at least one primary shard for the index [.security-7] is unavailable"} {"type":"response","@timestamp":"2020-04-28T03:27:38Z","tags":[],"pid":6,"method":"post","statusCode":500,"req":{"url":"/api/telemetry/v2/clusters/_stats","method":"post","headers":{"host":"test-telemetry-mfln-kb-http.e2e-b84wj-mercury.svc:5601","user-agent":"Go-http-client/1.1","content-length":"50","content-type":"application/json","kbn-version":"7.7.0-SNAPSHOT","accept-encoding":"gzip"},"remoteAddress":"10.113.34.8","userAgent":"10.113.34.8"},"res":{"statusCode":500,"responseTime":97,"contentLength":9},"message":"POST /api/telemetry/v2/clusters/_stats 500 97ms - 9.0B"} ``` Also reported in Elasticsearch logs: ``` {"type": "server", "timestamp": "2020-04-28T03:27:38,776Z", "level": "DEBUG", "component": "o.e.c.s.MasterService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "took [0s] to notify listeners on successful publication of cluster state (version: 47, uuid: omA1LLSVQruXPbYIVMw_Sw) for [create-index [.security-7], cause [api]]", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } {"type": "server", "timestamp": "2020-04-28T03:27:38,882Z", "level": "WARN", "component": "r.suppressed", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "path: /_security/user/_has_privileges, params: {}", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" , "stacktrace": ["org.elasticsearch.action.UnavailableShardsException: at least one primary shard for the index [.security-7] is unavailable", "at org.elasticsearch.xpack.security.support.SecurityIndexManager.getUnavailableReason(SecurityIndexManager.java:181) ~[x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.store.NativePrivilegeStore.getPrivileges(NativePrivilegeStore.java:102) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.user.TransportHasPrivilegesAction.resolveApplicationPrivileges(TransportHasPrivilegesAction.java:88) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.user.TransportHasPrivilegesAction.doExecute(TransportHasPrivilegesAction.java:80) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.user.TransportHasPrivilegesAction.doExecute(TransportHasPrivilegesAction.java:37) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:153) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$0(SecurityActionFilter.java:86) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$authorizeRequest$4(SecurityActionFilter.java:172) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorizeAction$4(AuthorizationService.java:247) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:625) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:600) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.authorizeClusterAction(RBACEngine.java:147) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.authorizeAction(AuthorizationService.java:249) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.maybeAuthorizeRunAs(AuthorizationService.java:232) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorize$1(AuthorizationService.java:197) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.lambda$resolveAuthorizationInfo$1(RBACEngine.java:120) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.store.CompositeRolesStore.getRoles(CompositeRolesStore.java:249) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.getRoles(RBACEngine.java:126) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.resolveAuthorizationInfo(RBACEngine.java:114) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.authorize(AuthorizationService.java:199) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.authorizeRequest(SecurityActionFilter.java:172) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$applyInternal$3(SecurityActionFilter.java:158) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:324) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:386) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:397) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:321) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:263) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:157) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.applyInternal(SecurityActionFilter.java:155) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.apply(SecurityActionFilter.java:107) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:151) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:129) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:64) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.client.node.NodeClient.executeLocally(NodeClient.java:83) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.client.node.NodeClient.doExecute(NodeClient.java:72) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:399) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionRequestBuilder.execute(ActionRequestBuilder.java:70) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.action.user.RestHasPrivilegesAction.lambda$innerPrepareRequest$1(RestHasPrivilegesAction.java:88) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.BaseRestHandler.handleRequest(BaseRestHandler.java:115) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.SecurityRestFilter.lambda$handleRequest$0(SecurityRestFilter.java:74) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.SecondaryAuthenticator.lambda$authenticateAndAttachToContext$2(SecondaryAuthenticator.java:82) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.SecondaryAuthenticator.authenticate(SecondaryAuthenticator.java:92) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.SecondaryAuthenticator.authenticateAndAttachToContext(SecondaryAuthenticator.java:77) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.SecurityRestFilter.lambda$handleRequest$2(SecurityRestFilter.java:68) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$writeAuthToContext$24(AuthenticationService.java:673) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.writeAuthToContext(AuthenticationService.java:687) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.finishAuthentication(AuthenticationService.java:661) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.consumeUser(AuthenticationService.java:608) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$consumeToken$16(AuthenticationService.java:491) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.core.common.IteratingActionListener.onResponse(IteratingActionListener.java:120) [x-pack-core-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$consumeToken$13(AuthenticationService.java:461) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.lambda$authenticateWithCache$1(CachingUsernamePasswordRealm.java:146) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.handleCachedAuthentication(CachingUsernamePasswordRealm.java:197) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.lambda$authenticateWithCache$2(CachingUsernamePasswordRealm.java:138) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.ListenableFuture$1.doRun(ListenableFuture.java:112) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:225) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.ListenableFuture.notifyListener(ListenableFuture.java:106) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.ListenableFuture.addListener(ListenableFuture.java:68) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.authenticateWithCache(CachingUsernamePasswordRealm.java:133) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.authenticate(CachingUsernamePasswordRealm.java:104) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$consumeToken$15(AuthenticationService.java:450) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.core.common.IteratingActionListener.run(IteratingActionListener.java:102) [x-pack-core-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.consumeToken(AuthenticationService.java:495) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$extractToken$11(AuthenticationService.java:417) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.extractToken(AuthenticationService.java:427) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$checkForApiKey$3(AuthenticationService.java:368) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.ApiKeyService.authenticateWithApiKeyIfPresent(ApiKeyService.java:321) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.checkForApiKey(AuthenticationService.java:348) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$0(AuthenticationService.java:330) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.TokenService.getAndValidateToken(TokenService.java:396) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:326) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:386) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:397) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:321) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:263) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:141) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:126) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.SecurityRestFilter.handleRequest(SecurityRestFilter.java:61) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.RestController.dispatchRequest(RestController.java:236) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.RestController.tryAllHandlers(RestController.java:318) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.RestController.dispatchRequest(RestController.java:176) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.AbstractHttpServerTransport.dispatchRequest(AbstractHttpServerTransport.java:329) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.AbstractHttpServerTransport.handleIncomingRequest(AbstractHttpServerTransport.java:383) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.AbstractHttpServerTransport.incomingRequest(AbstractHttpServerTransport.java:308) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.netty4.Netty4HttpRequestHandler.channelRead0(Netty4HttpRequestHandler.java:54) [transport-netty4-client-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.netty4.Netty4HttpRequestHandler.channelRead0(Netty4HttpRequestHandler.java:29) [transport-netty4-client-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at org.elasticsearch.http.netty4.Netty4HttpPipeliningHandler.channelRead(Netty4HttpPipeliningHandler.java:58) [transport-netty4-client-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1470) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1219) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1266) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:498) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:437) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:276) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) [netty-common-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.45.Final.jar:4.1.45.Final]", "at java.lang.Thread.run(Thread.java:832) [?:?]"] } {"type": "server", "timestamp": "2020-04-28T03:27:39,043Z", "level": "TRACE", "component": "o.e.c.s.MasterService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "will process [shard-started StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}[StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}]]", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } {"type": "server", "timestamp": "2020-04-28T03:27:39,043Z", "level": "DEBUG", "component": "o.e.c.s.MasterService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "executing cluster state update for [shard-started StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}[StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}]]", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } {"type": "server", "timestamp": "2020-04-28T03:27:39,044Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.security-7][0]]]).", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } ```
1.0
Error 500 in E2E test TestTelemetry - https://devops-ci.elastic.co/job/cloud-on-k8s-e2e-tests-snapshot-versions/42/testReport/github/com_elastic_cloud-on-k8s_test_e2e_kb/TestTelemetry_Kibana_should_expose_eck_info_in_telemetry_data/ ``` === RUN TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data: telemetry_test.go:46: Error Trace: telemetry_test.go:46 Error: Received unexpected error: fail to request /api/telemetry/v2/clusters/_stats, status is 500) Test: TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data {"log.level":"error","@timestamp":"2020-04-28T03:27:38.944Z","message":"stopping early","service.version":"0.0.0-00000000","service.type":"eck","ecs.version":"1.4.0","error":"test failure","error.stack_trace":"github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/pkg/mod/github.com/go-logr/zapr@v0.1.0/zapr.go:128\ngithub.com/elastic/cloud-on-k8s/test/e2e/test.StepList.RunSequential\n\t/go/src/github.com/elastic/cloud-on-k8s/test/e2e/test/step.go:43\ngithub.com/elastic/cloud-on-k8s/test/e2e/kb.TestTelemetry\n\t/go/src/github.com/elastic/cloud-on-k8s/test/e2e/kb/telemetry_test.go:60\ntesting.tRunner\n\t/usr/local/go/src/testing/testing.go:991"} --- FAIL: TestTelemetry/Kibana_should_expose_eck_info_in_telemetry_data (0.13s) ``` Kibana logs report a failure on the `.security_7` shard: ``` {"type":"error","@timestamp":"2020-04-28T03:27:38Z","tags":[],"pid":6,"level":"error","error":{"message":"[unavailable_shards_exception] at least one primary shard for the index [.security-7] is unavailable","name":"Error","stack":"Error: [unavailable_shards_exception] at least one primary shard for the index [.security-7] is unavailable\n at respond (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:349:15)\n at checkRespForFailure (/usr/share/kibana/node_modules/elasticsearch/src/lib/transport.js:306:7)\n at HttpConnector.<anonymous> (/usr/share/kibana/node_modules/elasticsearch/src/lib/connectors/http.js:173:7)\n at IncomingMessage.wrapper (/usr/share/kibana/node_modules/elasticsearch/node_modules/lodash/lodash.js:4929:19)\n at IncomingMessage.emit (events.js:203:15)\n at endReadableNT (_stream_readable.js:1145:12)\n at process._tickCallback (internal/process/next_tick.js:63:19)"},"url":{"protocol":null,"slashes":null,"auth":null,"host":null,"port":null,"hostname":null,"hash":null,"search":null,"query":{},"pathname":"/api/telemetry/v2/clusters/_stats","path":"/api/telemetry/v2/clusters/_stats","href":"/api/telemetry/v2/clusters/_stats"},"message":"[unavailable_shards_exception] at least one primary shard for the index [.security-7] is unavailable"} {"type":"response","@timestamp":"2020-04-28T03:27:38Z","tags":[],"pid":6,"method":"post","statusCode":500,"req":{"url":"/api/telemetry/v2/clusters/_stats","method":"post","headers":{"host":"test-telemetry-mfln-kb-http.e2e-b84wj-mercury.svc:5601","user-agent":"Go-http-client/1.1","content-length":"50","content-type":"application/json","kbn-version":"7.7.0-SNAPSHOT","accept-encoding":"gzip"},"remoteAddress":"10.113.34.8","userAgent":"10.113.34.8"},"res":{"statusCode":500,"responseTime":97,"contentLength":9},"message":"POST /api/telemetry/v2/clusters/_stats 500 97ms - 9.0B"} ``` Also reported in Elasticsearch logs: ``` {"type": "server", "timestamp": "2020-04-28T03:27:38,776Z", "level": "DEBUG", "component": "o.e.c.s.MasterService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "took [0s] to notify listeners on successful publication of cluster state (version: 47, uuid: omA1LLSVQruXPbYIVMw_Sw) for [create-index [.security-7], cause [api]]", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } {"type": "server", "timestamp": "2020-04-28T03:27:38,882Z", "level": "WARN", "component": "r.suppressed", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "path: /_security/user/_has_privileges, params: {}", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" , "stacktrace": ["org.elasticsearch.action.UnavailableShardsException: at least one primary shard for the index [.security-7] is unavailable", "at org.elasticsearch.xpack.security.support.SecurityIndexManager.getUnavailableReason(SecurityIndexManager.java:181) ~[x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.store.NativePrivilegeStore.getPrivileges(NativePrivilegeStore.java:102) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.user.TransportHasPrivilegesAction.resolveApplicationPrivileges(TransportHasPrivilegesAction.java:88) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.user.TransportHasPrivilegesAction.doExecute(TransportHasPrivilegesAction.java:80) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.user.TransportHasPrivilegesAction.doExecute(TransportHasPrivilegesAction.java:37) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:153) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$0(SecurityActionFilter.java:86) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$authorizeRequest$4(SecurityActionFilter.java:172) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorizeAction$4(AuthorizationService.java:247) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:625) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:600) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.authorizeClusterAction(RBACEngine.java:147) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.authorizeAction(AuthorizationService.java:249) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.maybeAuthorizeRunAs(AuthorizationService.java:232) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorize$1(AuthorizationService.java:197) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.lambda$resolveAuthorizationInfo$1(RBACEngine.java:120) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.store.CompositeRolesStore.getRoles(CompositeRolesStore.java:249) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.getRoles(RBACEngine.java:126) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.RBACEngine.resolveAuthorizationInfo(RBACEngine.java:114) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authz.AuthorizationService.authorize(AuthorizationService.java:199) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.authorizeRequest(SecurityActionFilter.java:172) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$applyInternal$3(SecurityActionFilter.java:158) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:324) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:386) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:397) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:321) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:263) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:157) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.applyInternal(SecurityActionFilter.java:155) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.apply(SecurityActionFilter.java:107) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:151) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:129) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:64) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.client.node.NodeClient.executeLocally(NodeClient.java:83) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.client.node.NodeClient.doExecute(NodeClient.java:72) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:399) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionRequestBuilder.execute(ActionRequestBuilder.java:70) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.action.user.RestHasPrivilegesAction.lambda$innerPrepareRequest$1(RestHasPrivilegesAction.java:88) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.BaseRestHandler.handleRequest(BaseRestHandler.java:115) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.SecurityRestFilter.lambda$handleRequest$0(SecurityRestFilter.java:74) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.SecondaryAuthenticator.lambda$authenticateAndAttachToContext$2(SecondaryAuthenticator.java:82) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.SecondaryAuthenticator.authenticate(SecondaryAuthenticator.java:92) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.SecondaryAuthenticator.authenticateAndAttachToContext(SecondaryAuthenticator.java:77) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.SecurityRestFilter.lambda$handleRequest$2(SecurityRestFilter.java:68) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$writeAuthToContext$24(AuthenticationService.java:673) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.writeAuthToContext(AuthenticationService.java:687) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.finishAuthentication(AuthenticationService.java:661) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.consumeUser(AuthenticationService.java:608) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$consumeToken$16(AuthenticationService.java:491) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.core.common.IteratingActionListener.onResponse(IteratingActionListener.java:120) [x-pack-core-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$consumeToken$13(AuthenticationService.java:461) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.lambda$authenticateWithCache$1(CachingUsernamePasswordRealm.java:146) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.handleCachedAuthentication(CachingUsernamePasswordRealm.java:197) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.lambda$authenticateWithCache$2(CachingUsernamePasswordRealm.java:138) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.ListenableFuture$1.doRun(ListenableFuture.java:112) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:225) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.ListenableFuture.notifyListener(ListenableFuture.java:106) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.common.util.concurrent.ListenableFuture.addListener(ListenableFuture.java:68) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.authenticateWithCache(CachingUsernamePasswordRealm.java:133) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.support.CachingUsernamePasswordRealm.authenticate(CachingUsernamePasswordRealm.java:104) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$consumeToken$15(AuthenticationService.java:450) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.core.common.IteratingActionListener.run(IteratingActionListener.java:102) [x-pack-core-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.consumeToken(AuthenticationService.java:495) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$extractToken$11(AuthenticationService.java:417) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.extractToken(AuthenticationService.java:427) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$checkForApiKey$3(AuthenticationService.java:368) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.ApiKeyService.authenticateWithApiKeyIfPresent(ApiKeyService.java:321) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.checkForApiKey(AuthenticationService.java:348) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$0(AuthenticationService.java:330) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:63) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.TokenService.getAndValidateToken(TokenService.java:396) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:326) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:386) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:397) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:321) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:263) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:141) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:126) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.xpack.security.rest.SecurityRestFilter.handleRequest(SecurityRestFilter.java:61) [x-pack-security-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.RestController.dispatchRequest(RestController.java:236) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.RestController.tryAllHandlers(RestController.java:318) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.rest.RestController.dispatchRequest(RestController.java:176) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.AbstractHttpServerTransport.dispatchRequest(AbstractHttpServerTransport.java:329) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.AbstractHttpServerTransport.handleIncomingRequest(AbstractHttpServerTransport.java:383) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.AbstractHttpServerTransport.incomingRequest(AbstractHttpServerTransport.java:308) [elasticsearch-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.netty4.Netty4HttpRequestHandler.channelRead0(Netty4HttpRequestHandler.java:54) [transport-netty4-client-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at org.elasticsearch.http.netty4.Netty4HttpRequestHandler.channelRead0(Netty4HttpRequestHandler.java:29) [transport-netty4-client-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:99) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at org.elasticsearch.http.netty4.Netty4HttpPipeliningHandler.channelRead(Netty4HttpPipeliningHandler.java:58) [transport-netty4-client-7.7.0-SNAPSHOT.jar:7.7.0-SNAPSHOT]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1470) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1219) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1266) [netty-handler-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:498) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:437) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:276) [netty-codec-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) [netty-transport-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) [netty-common-4.1.45.Final.jar:4.1.45.Final]", "at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.45.Final.jar:4.1.45.Final]", "at java.lang.Thread.run(Thread.java:832) [?:?]"] } {"type": "server", "timestamp": "2020-04-28T03:27:39,043Z", "level": "TRACE", "component": "o.e.c.s.MasterService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "will process [shard-started StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}[StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}]]", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } {"type": "server", "timestamp": "2020-04-28T03:27:39,043Z", "level": "DEBUG", "component": "o.e.c.s.MasterService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "executing cluster state update for [shard-started StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}[StartedShardEntry{shardId [[.security-7][0]], allocationId [-JRi0rd-RtStuBXpWN5sgQ], primary term [1], message [after new shard recovery]}]]", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } {"type": "server", "timestamp": "2020-04-28T03:27:39,044Z", "level": "INFO", "component": "o.e.c.r.a.AllocationService", "cluster.name": "test-telemetry-js26", "node.name": "test-telemetry-js26-es-masterdata-0", "message": "Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.security-7][0]]]).", "cluster.uuid": "6rUaPbbmSZCpxuRq6_hw1w", "node.id": "ZvkNQgTlToKjUVwVeZbmAw" } ```
test
error in test testtelemetry run testtelemetry kibana should expose eck info in telemetry data testtelemetry kibana should expose eck info in telemetry data telemetry test go error trace telemetry test go error received unexpected error fail to request api telemetry clusters stats status is test testtelemetry kibana should expose eck info in telemetry data log level error timestamp message stopping early service version service type eck ecs version error test failure error stack trace github com go logr zapr zaplogger error n t go pkg mod github com go logr zapr zapr go ngithub com elastic cloud on test test steplist runsequential n t go src github com elastic cloud on test test step go ngithub com elastic cloud on test kb testtelemetry n t go src github com elastic cloud on test kb telemetry test go ntesting trunner n t usr local go src testing testing go fail testtelemetry kibana should expose eck info in telemetry data kibana logs report a failure on the security shard type error timestamp tags pid level error error message at least one primary shard for the index is unavailable name error stack error at least one primary shard for the index is unavailable n at respond usr share kibana node modules elasticsearch src lib transport js n at checkrespforfailure usr share kibana node modules elasticsearch src lib transport js n at httpconnector usr share kibana node modules elasticsearch src lib connectors http js n at incomingmessage wrapper usr share kibana node modules elasticsearch node modules lodash lodash js n at incomingmessage emit events js n at endreadablent stream readable js n at process tickcallback internal process next tick js url protocol null slashes null auth null host null port null hostname null hash null search null query pathname api telemetry clusters stats path api telemetry clusters stats href api telemetry clusters stats message at least one primary shard for the index is unavailable type response timestamp tags pid method post statuscode req url api telemetry clusters stats method post headers host test telemetry mfln kb http mercury svc user agent go http client content length content type application json kbn version snapshot accept encoding gzip remoteaddress useragent res statuscode responsetime contentlength message post api telemetry clusters stats also reported in elasticsearch logs type server timestamp level debug component o e c s masterservice cluster name test telemetry node name test telemetry es masterdata message took to notify listeners on successful publication of cluster state version uuid sw for cause cluster uuid node id zvknqgtltokjuvwvezbmaw type server timestamp level warn component r suppressed cluster name test telemetry node name test telemetry es masterdata message path security user has privileges params cluster uuid node id zvknqgtltokjuvwvezbmaw stacktrace is unavailable at org elasticsearch xpack security support securityindexmanager getunavailablereason securityindexmanager java at org elasticsearch xpack security authz store nativeprivilegestore getprivileges nativeprivilegestore java at org elasticsearch xpack security action user transporthasprivilegesaction resolveapplicationprivileges transporthasprivilegesaction java at org elasticsearch xpack security action user transporthasprivilegesaction doexecute transporthasprivilegesaction java at org elasticsearch xpack security action user transporthasprivilegesaction doexecute transporthasprivilegesaction java at org elasticsearch action support transportaction requestfilterchain proceed transportaction java at org elasticsearch xpack security action filter securityactionfilter lambda apply securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security action filter securityactionfilter lambda authorizerequest securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authz authorizationservice lambda authorizeaction authorizationservice java at org elasticsearch xpack security authz authorizationservice authorizationresultlistener onresponse authorizationservice java at org elasticsearch xpack security authz authorizationservice authorizationresultlistener onresponse authorizationservice java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch xpack security authz rbacengine authorizeclusteraction rbacengine java at org elasticsearch xpack security authz authorizationservice authorizeaction authorizationservice java at org elasticsearch xpack security authz authorizationservice maybeauthorizerunas authorizationservice java at org elasticsearch xpack security authz authorizationservice lambda authorize authorizationservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch xpack security authz rbacengine lambda resolveauthorizationinfo rbacengine java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authz store compositerolesstore getroles compositerolesstore java at org elasticsearch xpack security authz rbacengine getroles rbacengine java at org elasticsearch xpack security authz rbacengine resolveauthorizationinfo rbacengine java at org elasticsearch xpack security authz authorizationservice authorize authorizationservice java at org elasticsearch xpack security action filter securityactionfilter authorizerequest securityactionfilter java at org elasticsearch xpack security action filter securityactionfilter lambda applyinternal securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc authenticationservice authenticator lambda authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator access authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticate authenticationservice java at org elasticsearch xpack security action filter securityactionfilter applyinternal securityactionfilter java at org elasticsearch xpack security action filter securityactionfilter apply securityactionfilter java at org elasticsearch action support transportaction requestfilterchain proceed transportaction java at org elasticsearch action support transportaction execute transportaction java at org elasticsearch action support transportaction execute transportaction java at org elasticsearch client node nodeclient executelocally nodeclient java at org elasticsearch client node nodeclient doexecute nodeclient java at org elasticsearch client support abstractclient execute abstractclient java at org elasticsearch action actionrequestbuilder execute actionrequestbuilder java at org elasticsearch xpack security rest action user resthasprivilegesaction lambda innerpreparerequest resthasprivilegesaction java at org elasticsearch rest baseresthandler handlerequest baseresthandler java at org elasticsearch xpack security rest securityrestfilter lambda handlerequest securityrestfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc support secondaryauthenticator lambda authenticateandattachtocontext secondaryauthenticator java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc support secondaryauthenticator authenticate secondaryauthenticator java at org elasticsearch xpack security authc support secondaryauthenticator authenticateandattachtocontext secondaryauthenticator java at org elasticsearch xpack security rest securityrestfilter lambda handlerequest securityrestfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc authenticationservice authenticator lambda writeauthtocontext authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator writeauthtocontext authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator finishauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator consumeuser authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda consumetoken authenticationservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch xpack core common iteratingactionlistener onresponse iteratingactionlistener java at org elasticsearch xpack security authc authenticationservice authenticator lambda consumetoken authenticationservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc support cachingusernamepasswordrealm lambda authenticatewithcache cachingusernamepasswordrealm java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc support cachingusernamepasswordrealm handlecachedauthentication cachingusernamepasswordrealm java at org elasticsearch xpack security authc support cachingusernamepasswordrealm lambda authenticatewithcache cachingusernamepasswordrealm java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch common util concurrent listenablefuture dorun listenablefuture java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java at org elasticsearch common util concurrent esexecutors directexecutorservice execute esexecutors java at org elasticsearch common util concurrent listenablefuture notifylistener listenablefuture java at org elasticsearch common util concurrent listenablefuture addlistener listenablefuture java at org elasticsearch xpack security authc support cachingusernamepasswordrealm authenticatewithcache cachingusernamepasswordrealm java at org elasticsearch xpack security authc support cachingusernamepasswordrealm authenticate cachingusernamepasswordrealm java at org elasticsearch xpack security authc authenticationservice authenticator lambda consumetoken authenticationservice java at org elasticsearch xpack core common iteratingactionlistener run iteratingactionlistener java at org elasticsearch xpack security authc authenticationservice authenticator consumetoken authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda extracttoken authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator extracttoken authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda checkforapikey authenticationservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc apikeyservice authenticatewithapikeyifpresent apikeyservice java at org elasticsearch xpack security authc authenticationservice authenticator checkforapikey authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda authenticateasync authenticationservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc tokenservice getandvalidatetoken tokenservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator access authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticate authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticate authenticationservice java at org elasticsearch xpack security rest securityrestfilter handlerequest securityrestfilter java at org elasticsearch rest restcontroller dispatchrequest restcontroller java at org elasticsearch rest restcontroller tryallhandlers restcontroller java at org elasticsearch rest restcontroller dispatchrequest restcontroller java at org elasticsearch http abstracthttpservertransport dispatchrequest abstracthttpservertransport java at org elasticsearch http abstracthttpservertransport handleincomingrequest abstracthttpservertransport java at org elasticsearch http abstracthttpservertransport incomingrequest abstracthttpservertransport java at org elasticsearch http java at org elasticsearch http java at io netty channel simplechannelinboundhandler channelread simplechannelinboundhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at org elasticsearch http channelread java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler timeout idlestatehandler channelread idlestatehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler ssl sslhandler unwrap sslhandler java at io netty handler ssl sslhandler decodejdkcompatible sslhandler java at io netty handler ssl sslhandler decode sslhandler java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysplain nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at java lang thread run thread java type server timestamp level trace component o e c s masterservice cluster name test telemetry node name test telemetry es masterdata message will process allocationid primary term message allocationid primary term message cluster uuid node id zvknqgtltokjuvwvezbmaw type server timestamp level debug component o e c s masterservice cluster name test telemetry node name test telemetry es masterdata message executing cluster state update for allocationid primary term message allocationid primary term message cluster uuid node id zvknqgtltokjuvwvezbmaw type server timestamp level info component o e c r a allocationservice cluster name test telemetry node name test telemetry es masterdata message cluster health status changed from to reason cluster uuid node id zvknqgtltokjuvwvezbmaw
1
225,426
17,858,642,579
IssuesEvent
2021-09-05 14:31:26
vladimirdimov99/Teodor.bg
https://api.github.com/repos/vladimirdimov99/Teodor.bg
opened
Suite - Authorization - Sign Up, Test ID - TEODOR - 005, Name - Sign Up with valid credentials
positive test case
**Description** Validating that user can sign up to the website with valid credentials **Author** Vladimir Dimov **Priority** High **Behavior** Positive **Type** Functional **Preconditions** 1. Open https://teodor.bg/ 2. Click on the profile icon at the top right. 3. Click on the “Sign Up“ button. **Steps to reproduce** 1. Type valid credentials in the required input fields. **Expected result** Credentials are accepted 2. Tick the required options. **Expected result** Required options are ticked 3. Click on the “Sign Up“ button. **Expected result** User signed up successfully to the website
1.0
Suite - Authorization - Sign Up, Test ID - TEODOR - 005, Name - Sign Up with valid credentials - **Description** Validating that user can sign up to the website with valid credentials **Author** Vladimir Dimov **Priority** High **Behavior** Positive **Type** Functional **Preconditions** 1. Open https://teodor.bg/ 2. Click on the profile icon at the top right. 3. Click on the “Sign Up“ button. **Steps to reproduce** 1. Type valid credentials in the required input fields. **Expected result** Credentials are accepted 2. Tick the required options. **Expected result** Required options are ticked 3. Click on the “Sign Up“ button. **Expected result** User signed up successfully to the website
test
suite authorization sign up test id teodor name sign up with valid credentials description validating that user can sign up to the website with valid credentials author vladimir dimov priority high behavior positive type functional preconditions open click on the profile icon at the top right click on the “sign up“ button steps to reproduce type valid credentials in the required input fields expected result credentials are accepted tick the required options expected result required options are ticked click on the “sign up“ button expected result user signed up successfully to the website
1
57,796
6,557,527,359
IssuesEvent
2017-09-06 17:46:30
ESMCI/cime
https://api.github.com/repos/ESMCI/cime
closed
ERI test's cprnc.out files overwrite each other
Low Priority st: invalid tp: system tests ty: Bug
I don't think this causes any significant problems, but makes debugging a failed test harder: An ERI test does two comparisons at the end of the test: ```python self._component_compare_test("base", "hybrid") self._component_compare_test("base", "rest") ``` However, due to #453 (closed as a wontfix), both of these sets of cprnc.out files are written with a `.base` suffix, leading the second set to overwrite the first. Solutions could be either: 1. Resolving #453 - going back to the old behavior of using the second suffix for the suffix of the cprnc.out file 2. Changing one of the comparisons in the eri test; for example, I'd suggest: ```python self._component_compare_test("hybrid", "base") self._component_compare_test("base", "rest") ``` What do others think? Should we just do the simple solution of (2) for now? Or could this have any unforeseen consequences?
1.0
ERI test's cprnc.out files overwrite each other - I don't think this causes any significant problems, but makes debugging a failed test harder: An ERI test does two comparisons at the end of the test: ```python self._component_compare_test("base", "hybrid") self._component_compare_test("base", "rest") ``` However, due to #453 (closed as a wontfix), both of these sets of cprnc.out files are written with a `.base` suffix, leading the second set to overwrite the first. Solutions could be either: 1. Resolving #453 - going back to the old behavior of using the second suffix for the suffix of the cprnc.out file 2. Changing one of the comparisons in the eri test; for example, I'd suggest: ```python self._component_compare_test("hybrid", "base") self._component_compare_test("base", "rest") ``` What do others think? Should we just do the simple solution of (2) for now? Or could this have any unforeseen consequences?
test
eri test s cprnc out files overwrite each other i don t think this causes any significant problems but makes debugging a failed test harder an eri test does two comparisons at the end of the test python self component compare test base hybrid self component compare test base rest however due to closed as a wontfix both of these sets of cprnc out files are written with a base suffix leading the second set to overwrite the first solutions could be either resolving going back to the old behavior of using the second suffix for the suffix of the cprnc out file changing one of the comparisons in the eri test for example i d suggest python self component compare test hybrid base self component compare test base rest what do others think should we just do the simple solution of for now or could this have any unforeseen consequences
1
174,690
6,542,402,932
IssuesEvent
2017-09-02 06:01:12
blynkkk/blynk-library
https://api.github.com/repos/blynkkk/blynk-library
closed
ButtonWidget for syncing physical and virtual
feature low priority
I think it's time now for a ButtonWidget class for blynk library that handles synchronization between a virtual and physical pin. Initialization and instantiation: ButtonWidget button1(virtualPin, physicalPin) Sync is handled in background by Blynk.run() Extra suggested functions: .pinMode(Mode) // just for altering depending on circuit .vibrate(boolean) // vibrates phone when physical button is pressed
1.0
ButtonWidget for syncing physical and virtual - I think it's time now for a ButtonWidget class for blynk library that handles synchronization between a virtual and physical pin. Initialization and instantiation: ButtonWidget button1(virtualPin, physicalPin) Sync is handled in background by Blynk.run() Extra suggested functions: .pinMode(Mode) // just for altering depending on circuit .vibrate(boolean) // vibrates phone when physical button is pressed
non_test
buttonwidget for syncing physical and virtual i think it s time now for a buttonwidget class for blynk library that handles synchronization between a virtual and physical pin initialization and instantiation buttonwidget virtualpin physicalpin sync is handled in background by blynk run extra suggested functions pinmode mode just for altering depending on circuit vibrate boolean vibrates phone when physical button is pressed
0
39,183
5,222,020,885
IssuesEvent
2017-01-27 05:31:22
jemalloc/jemalloc
https://api.github.com/repos/jemalloc/jemalloc
closed
Beef up CI matrix
testing
We should also test (at least) some permutations of --enable-debug, --disable-stats and --disable-tcache.
1.0
Beef up CI matrix - We should also test (at least) some permutations of --enable-debug, --disable-stats and --disable-tcache.
test
beef up ci matrix we should also test at least some permutations of enable debug disable stats and disable tcache
1
258,814
22,349,814,298
IssuesEvent
2022-06-15 11:00:51
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
opened
rv_core_Ibex - Double fault test
Type:Task IP:rv_core_ibex Component:ChipLevelTest
# Verify the functionality of the ibex double fault dump. - Purposely create an ibex double exception during execution, by performing an unmapped read and in the exception handler perform another unmapped read. - Ensure the rstmgr fault dump correctly captures both dumps correctly and indicates the previous dump is valid.
1.0
rv_core_Ibex - Double fault test - # Verify the functionality of the ibex double fault dump. - Purposely create an ibex double exception during execution, by performing an unmapped read and in the exception handler perform another unmapped read. - Ensure the rstmgr fault dump correctly captures both dumps correctly and indicates the previous dump is valid.
test
rv core ibex double fault test verify the functionality of the ibex double fault dump purposely create an ibex double exception during execution by performing an unmapped read and in the exception handler perform another unmapped read ensure the rstmgr fault dump correctly captures both dumps correctly and indicates the previous dump is valid
1
56,369
6,517,907,520
IssuesEvent
2017-08-28 04:27:33
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
[FVT] automate IB test
component:test priority:high sprint1 type:feature
what to do : * [ ] implement these test cases * enable_ib_in_diskful_rhels * enable_ib_in_diskless_rhels * confignetwork_ib_ipoib * [ ] Test these test cases in target environment * [ ] Replace confignics with confignetwork for ib configuration document
1.0
[FVT] automate IB test - what to do : * [ ] implement these test cases * enable_ib_in_diskful_rhels * enable_ib_in_diskless_rhels * confignetwork_ib_ipoib * [ ] Test these test cases in target environment * [ ] Replace confignics with confignetwork for ib configuration document
test
automate ib test what to do implement these test cases enable ib in diskful rhels enable ib in diskless rhels confignetwork ib ipoib test these test cases in target environment replace confignics with confignetwork for ib configuration document
1
166,433
12,955,464,334
IssuesEvent
2020-07-20 06:23:36
woowacourse-teams/2020-devbie
https://api.github.com/repos/woowacourse-teams/2020-devbie
closed
[TEST] JwtTokenProvider 테스트 코드 추가
🧪 Test
## 목적 및 필요성 - JwtTokenProvider에 대한 테스트 코드가 필요하다. ## 작업 상세 내용 - [ ] JwtTokenProviderTest 테스트 클래스를 만들어 필요한 단위 테스트를 작성한다.
1.0
[TEST] JwtTokenProvider 테스트 코드 추가 - ## 목적 및 필요성 - JwtTokenProvider에 대한 테스트 코드가 필요하다. ## 작업 상세 내용 - [ ] JwtTokenProviderTest 테스트 클래스를 만들어 필요한 단위 테스트를 작성한다.
test
jwttokenprovider 테스트 코드 추가 목적 및 필요성 jwttokenprovider에 대한 테스트 코드가 필요하다 작업 상세 내용 jwttokenprovidertest 테스트 클래스를 만들어 필요한 단위 테스트를 작성한다
1
330,300
28,367,639,096
IssuesEvent
2023-04-12 14:50:02
nrwl/nx
https://api.github.com/repos/nrwl/nx
closed
Code coverage is wrong for sibling libraries
type: bug scope: testing tools
## Current Behavior Not sure if this is jest or nx related, but here it goes... When I run jest to collect the code coverage using the following command: `nx run my-library:test --testFile=the-best.component.spec.ts --codeCoverage` I only get code coverage for the library where I'm running the command (in this case `my-library`). The problem is that I have inside `the-best.component.ts` an alias import to a service inside a different module (for example, like this: `import { ExternalService } from "@app/external-library"`, that I'm not mocking inside `the-best.component.spec.ts`. This means that if I debug the test on `the-best.component.spec.ts` I can hit a breakpoint on `ExternalService`. My assumption is that if I can hit a breakpoint inside `ExternalService`, then that file should have that line marked has covered by my test on `the-best.component.spec.ts` A quick diagram: ``` |- my-library | |- the-best.component.spec.ts | |- the-best.component.ts | |- external-library | |- external.service.ts ``` ## Expected Behavior When running code coverage for tests on `the-best.component.spec.ts` I'd like to see code on `external.service.ts` to be covered if I'm not mocking it on `the-best.component.spec.ts`. ## Steps to Reproduce I believe you just need to create an nx project, have 2 libraries, and from one library call a service from the other library. In the unit tests on the first library don't mock the second library. ### Environment ``` Node : 12.22.10 OS : darwin x64 npm : 6.14.16 nx : Not Found @nrwl/angular : 12.10.1 @nrwl/cli : 12.10.1 @nrwl/cypress : 12.10.1 @nrwl/devkit : 12.10.1 @nrwl/eslint-plugin-nx : 12.10.0 @nrwl/express : Not Found @nrwl/jest : 12.10.1 @nrwl/linter : 12.10.1 @nrwl/nest : Not Found @nrwl/next : Not Found @nrwl/node : 12.10.1 @nrwl/nx-cloud : Not Found @nrwl/react : Not Found @nrwl/schematics : Not Found @nrwl/tao : 12.10.1 @nrwl/web : Not Found @nrwl/workspace : 12.10.1 @nrwl/storybook : 12.10.1 @nrwl/gatsby : Not Found typescript : 4.3.5 ```
1.0
Code coverage is wrong for sibling libraries - ## Current Behavior Not sure if this is jest or nx related, but here it goes... When I run jest to collect the code coverage using the following command: `nx run my-library:test --testFile=the-best.component.spec.ts --codeCoverage` I only get code coverage for the library where I'm running the command (in this case `my-library`). The problem is that I have inside `the-best.component.ts` an alias import to a service inside a different module (for example, like this: `import { ExternalService } from "@app/external-library"`, that I'm not mocking inside `the-best.component.spec.ts`. This means that if I debug the test on `the-best.component.spec.ts` I can hit a breakpoint on `ExternalService`. My assumption is that if I can hit a breakpoint inside `ExternalService`, then that file should have that line marked has covered by my test on `the-best.component.spec.ts` A quick diagram: ``` |- my-library | |- the-best.component.spec.ts | |- the-best.component.ts | |- external-library | |- external.service.ts ``` ## Expected Behavior When running code coverage for tests on `the-best.component.spec.ts` I'd like to see code on `external.service.ts` to be covered if I'm not mocking it on `the-best.component.spec.ts`. ## Steps to Reproduce I believe you just need to create an nx project, have 2 libraries, and from one library call a service from the other library. In the unit tests on the first library don't mock the second library. ### Environment ``` Node : 12.22.10 OS : darwin x64 npm : 6.14.16 nx : Not Found @nrwl/angular : 12.10.1 @nrwl/cli : 12.10.1 @nrwl/cypress : 12.10.1 @nrwl/devkit : 12.10.1 @nrwl/eslint-plugin-nx : 12.10.0 @nrwl/express : Not Found @nrwl/jest : 12.10.1 @nrwl/linter : 12.10.1 @nrwl/nest : Not Found @nrwl/next : Not Found @nrwl/node : 12.10.1 @nrwl/nx-cloud : Not Found @nrwl/react : Not Found @nrwl/schematics : Not Found @nrwl/tao : 12.10.1 @nrwl/web : Not Found @nrwl/workspace : 12.10.1 @nrwl/storybook : 12.10.1 @nrwl/gatsby : Not Found typescript : 4.3.5 ```
test
code coverage is wrong for sibling libraries current behavior not sure if this is jest or nx related but here it goes when i run jest to collect the code coverage using the following command nx run my library test testfile the best component spec ts codecoverage i only get code coverage for the library where i m running the command in this case my library the problem is that i have inside the best component ts an alias import to a service inside a different module for example like this import externalservice from app external library that i m not mocking inside the best component spec ts this means that if i debug the test on the best component spec ts i can hit a breakpoint on externalservice my assumption is that if i can hit a breakpoint inside externalservice then that file should have that line marked has covered by my test on the best component spec ts a quick diagram my library the best component spec ts the best component ts external library external service ts expected behavior when running code coverage for tests on the best component spec ts i d like to see code on external service ts to be covered if i m not mocking it on the best component spec ts steps to reproduce i believe you just need to create an nx project have libraries and from one library call a service from the other library in the unit tests on the first library don t mock the second library environment node os darwin npm nx not found nrwl angular nrwl cli nrwl cypress nrwl devkit nrwl eslint plugin nx nrwl express not found nrwl jest nrwl linter nrwl nest not found nrwl next not found nrwl node nrwl nx cloud not found nrwl react not found nrwl schematics not found nrwl tao nrwl web not found nrwl workspace nrwl storybook nrwl gatsby not found typescript
1
64,275
12,425,732,248
IssuesEvent
2020-05-24 17:42:30
Pokecube-Development/Pokecube-Issues-and-Wiki
https://api.github.com/repos/Pokecube-Development/Pokecube-Issues-and-Wiki
closed
gene extractor delete both the bottle and pokemon
1.14.x 1.15.2 Bug - Code Fixed
#### Issue Description: using the gene extractor to extract a certain gene from a pokemon deletes the bottle and pokemon #### What happens: using the gene extractor to extract a certain gene from a pokemon deletes the bottle and pokemon leaving you with an empty pokecube #### What you expected to happen: #### Steps to reproduce: 1.put pokemon, bottle, and selector in gene extractor 2.wait for it to finish 3.notice bottle and pokemon have been deleted ... ____ #### Affected Versions (Do *not* use "latest"): Replace with a list of all mods you have in. - Pokecube AIO:2.7.0 - Minecraft: 1.15.2 - Forge: 31.1.18
1.0
gene extractor delete both the bottle and pokemon - #### Issue Description: using the gene extractor to extract a certain gene from a pokemon deletes the bottle and pokemon #### What happens: using the gene extractor to extract a certain gene from a pokemon deletes the bottle and pokemon leaving you with an empty pokecube #### What you expected to happen: #### Steps to reproduce: 1.put pokemon, bottle, and selector in gene extractor 2.wait for it to finish 3.notice bottle and pokemon have been deleted ... ____ #### Affected Versions (Do *not* use "latest"): Replace with a list of all mods you have in. - Pokecube AIO:2.7.0 - Minecraft: 1.15.2 - Forge: 31.1.18
non_test
gene extractor delete both the bottle and pokemon issue description using the gene extractor to extract a certain gene from a pokemon deletes the bottle and pokemon what happens using the gene extractor to extract a certain gene from a pokemon deletes the bottle and pokemon leaving you with an empty pokecube what you expected to happen steps to reproduce put pokemon bottle and selector in gene extractor wait for it to finish notice bottle and pokemon have been deleted affected versions do not use latest replace with a list of all mods you have in pokecube aio minecraft forge
0
81,427
15,624,743,872
IssuesEvent
2021-03-21 04:17:19
kadirselcuk/matrix.to
https://api.github.com/repos/kadirselcuk/matrix.to
opened
CVE-2020-15138 (High) detected in prismjs-1.17.1.tgz
security vulnerability
## CVE-2020-15138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.17.1.tgz</b></p></summary> <p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p> <p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.17.1.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.17.1.tgz</a></p> <p>Path to dependency file: matrix.to/package.json</p> <p>Path to vulnerable library: matrix.to/node_modules/refractor/node_modules/prismjs/package.json</p> <p> Dependency Hierarchy: - addon-storysource-5.3.21.tgz (Root Library) - react-syntax-highlighter-11.0.2.tgz - refractor-2.10.1.tgz - :x: **prismjs-1.17.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kadirselcuk/matrix.to/commit/0e8a685a962d2458de04fd5153adb74f00f54e63">0e8a685a962d2458de04fd5153adb74f00f54e63</a></p> <p>Found in base branch: <b>matrixtwo/groups</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prism is vulnerable to Cross-Site Scripting. The easing preview of the Previewers plugin has an XSS vulnerability that allows attackers to execute arbitrary code in Safari and Internet Explorer. This impacts all Safari and Internet Explorer users of Prism >=v1.1.0 that use the _Previewers_ plugin (>=v1.10.0) or the _Previewer: Easing_ plugin (v1.1.0 to v1.9.0). This problem is fixed in version 1.21.0. To workaround the issue without upgrading, disable the easing preview on all impacted code blocks. You need Prism v1.10.0 or newer to apply this workaround. <p>Publish Date: 2020-08-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15138>CVE-2020-15138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/PrismJS/prism/v1.21.0">https://github.com/PrismJS/prism/v1.21.0</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: v1.21.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15138 (High) detected in prismjs-1.17.1.tgz - ## CVE-2020-15138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.17.1.tgz</b></p></summary> <p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p> <p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.17.1.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.17.1.tgz</a></p> <p>Path to dependency file: matrix.to/package.json</p> <p>Path to vulnerable library: matrix.to/node_modules/refractor/node_modules/prismjs/package.json</p> <p> Dependency Hierarchy: - addon-storysource-5.3.21.tgz (Root Library) - react-syntax-highlighter-11.0.2.tgz - refractor-2.10.1.tgz - :x: **prismjs-1.17.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kadirselcuk/matrix.to/commit/0e8a685a962d2458de04fd5153adb74f00f54e63">0e8a685a962d2458de04fd5153adb74f00f54e63</a></p> <p>Found in base branch: <b>matrixtwo/groups</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prism is vulnerable to Cross-Site Scripting. The easing preview of the Previewers plugin has an XSS vulnerability that allows attackers to execute arbitrary code in Safari and Internet Explorer. This impacts all Safari and Internet Explorer users of Prism >=v1.1.0 that use the _Previewers_ plugin (>=v1.10.0) or the _Previewer: Easing_ plugin (v1.1.0 to v1.9.0). This problem is fixed in version 1.21.0. To workaround the issue without upgrading, disable the easing preview on all impacted code blocks. You need Prism v1.10.0 or newer to apply this workaround. <p>Publish Date: 2020-08-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15138>CVE-2020-15138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/PrismJS/prism/v1.21.0">https://github.com/PrismJS/prism/v1.21.0</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: v1.21.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in prismjs tgz cve high severity vulnerability vulnerable library prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file matrix to package json path to vulnerable library matrix to node modules refractor node modules prismjs package json dependency hierarchy addon storysource tgz root library react syntax highlighter tgz refractor tgz x prismjs tgz vulnerable library found in head commit a href found in base branch matrixtwo groups vulnerability details prism is vulnerable to cross site scripting the easing preview of the previewers plugin has an xss vulnerability that allows attackers to execute arbitrary code in safari and internet explorer this impacts all safari and internet explorer users of prism that use the previewers plugin or the previewer easing plugin to this problem is fixed in version to workaround the issue without upgrading disable the easing preview on all impacted code blocks you need prism or newer to apply this workaround publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
101,444
4,117,848,109
IssuesEvent
2016-06-08 09:09:30
googlei18n/libphonenumber
https://api.github.com/repos/googlei18n/libphonenumber
closed
+60116XXXXXXXX should be a valid MY mobile phone number
bug priority-medium
Hello there, +60116XXXXXXXX should be a valid MY mobile phone number, but current lib returns as invalid.
1.0
+60116XXXXXXXX should be a valid MY mobile phone number - Hello there, +60116XXXXXXXX should be a valid MY mobile phone number, but current lib returns as invalid.
non_test
should be a valid my mobile phone number hello there should be a valid my mobile phone number but current lib returns as invalid
0
324,388
27,805,823,043
IssuesEvent
2023-03-17 19:47:59
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
closed
Validate prevent `wazuh-agents` from connecting to a lower `wazuh-manager` version
team/qa target/4.5.0 dev-testing role/qa-binary-beats subteam/qa-rainbow level/task type/test
| Target version | Related issue | Related PR | |--------------------|--------------------|-----------------| | 4.5 | [#11011](https://github.com/wazuh/wazuh/issues/11011) | [#15661](https://github.com/wazuh/wazuh/pull/15661) | <!-- Important: No section may be left blank. If not, delete it directly (in principle only Steps to reproduce could be left blank in case of not proceeding, although there are always exceptions). --> ## Description <!-- Description that puts into context and shows the QA tester the changes that have been made by the developer and need to be tested. --> With this modifications, the `wazuh-manager` wont allow an agent with higher version than it to connect, preventing compatibility issues. Although this was not supported before, there was no mechanism to avoid it. Also a new message is logged in agent `ossec.log` file informing this. ## Proposed checks <!-- Indicate through a list of checkboxes the suggested checks to be carried out by the QA tester --> > **Note:** To test some cases with an agent or manager v4.4 (or any version) with the new modification (the fixed code), its needed to create a modified 4.5 package where the version is manually lowered > <details><summary>How to create the modified package</summary> > > 1. Clone the branch with the fix `git clone https://github.com/wazuh/wazuh.git -b 11011-prevent-agents-from-lower-version` > 2. Modify the `__ossec_version` value in the file `src/headers/defs.h` (line 70) changing it to the desired version (it could be v4.4.0 or v4.3.6 or even v4.6.0 it doesnt matter) > 3. Now it could be installed by sources following [this guide](https://documentation.wazuh.com/current/deployment-options/wazuh-from-sources/wazuh-server/) or [this file](https://github.com/wazuh/wazuh/blob/11011-prevent-agents-from-lower-version/INSTALL), or a package could be generated with this modified code using the [package generation tool](https://documentation.wazuh.com/current/development/packaging/index.html) (using the flag `--sources <PATH>` indicating the path of this modified source instead of `-b <BRANCH>`) > </details> - <details><summary>Scenario outline 1: Enroll agents to wazuh-manager</summary> ```Gherkin Given a wazuh-manager "v4.5.0". And a <agent-OS> wazuh-agent "<agent-version>". When wazuh-agent tries to enroll to the wazuh-manager. Then the agent "<result>" to wazuh-manager. Examples: | agent-OS | agent-version | result | |----------|---------------|-----------------| | Ubuntu | v4.3.10 | connects | | Ubuntu | v4.4.0 | connects | | Ubuntu | v4.5.0 | connects | | Ubuntu | v4.6.0 | doesn't connect | | Macos | v4.3.10 | connects | | Macos | v4.4.0 | connects | | Macos | v4.5.0 | connects | | Macos | v4.6.0 | doesn't connect | | Windows | v4.3.10 | connects | | Windows | v4.4.0 | connects | | Windows | v4.5.0 | connects | | Windows | v4.6.0 | doesn't connect | ``` </details> - <details><summary>Scenario outline 2: Upgrade manager with agents connected</summary> ```Gherkin Given a wazuh-manager "<manager-initial-version>". And 10 wazuh-agents already connected. When wazuh-manager is upgraded to 4.5.0. Then the upgrade is successful without any error. Examples: | manager-initial-version | |-------------------------| | v4.0.0 | | v4.1.0 | | v4.2.0 | | v4.3.0 | | v4.4.0 | ``` </details> <!-- - <details><summary>Scenario outline 3: Upgrade agent connected to a manager</summary> ```Gherkin Given a wazuh-manager 4.5.0. And a wazuh-agent "<agent-initial-version>" already connected. When wazuh-agent is upgraded to "<agent-final-version>". Then the agent status is "<result>". Examples: | agent-OS | agent-initial-version | agent-final-version | result | |----------|-----------------------|-----------------------|-----------------------| | Ubuntu | v4.4.0 | v4.5.0 | Active | | Ubuntu | v4.5.0 | v4.6.0 | Disconnected | | Macos | v4.4.0 | v4.5.0 | Active | | Macos | v4.5.0 | v4.6.0 | Disconnected | | Windows | v4.4.0 | v4.5.0 | Active | | Windows | v4.5.0 | v4.6.0 | Disconnected | ``` </details> --> ## Expected results <!-- Indicate expected results such as behaviors, logs... --> The manager only accepts connections from agents of equal or lower versions than itself. ## Conclusions and considerations <!-- Indicate considerations to take into account when performing the testing that may not be very intuitive. If any configuration has to be applied, it should be indicated as far as possible. -->
2.0
Validate prevent `wazuh-agents` from connecting to a lower `wazuh-manager` version - | Target version | Related issue | Related PR | |--------------------|--------------------|-----------------| | 4.5 | [#11011](https://github.com/wazuh/wazuh/issues/11011) | [#15661](https://github.com/wazuh/wazuh/pull/15661) | <!-- Important: No section may be left blank. If not, delete it directly (in principle only Steps to reproduce could be left blank in case of not proceeding, although there are always exceptions). --> ## Description <!-- Description that puts into context and shows the QA tester the changes that have been made by the developer and need to be tested. --> With this modifications, the `wazuh-manager` wont allow an agent with higher version than it to connect, preventing compatibility issues. Although this was not supported before, there was no mechanism to avoid it. Also a new message is logged in agent `ossec.log` file informing this. ## Proposed checks <!-- Indicate through a list of checkboxes the suggested checks to be carried out by the QA tester --> > **Note:** To test some cases with an agent or manager v4.4 (or any version) with the new modification (the fixed code), its needed to create a modified 4.5 package where the version is manually lowered > <details><summary>How to create the modified package</summary> > > 1. Clone the branch with the fix `git clone https://github.com/wazuh/wazuh.git -b 11011-prevent-agents-from-lower-version` > 2. Modify the `__ossec_version` value in the file `src/headers/defs.h` (line 70) changing it to the desired version (it could be v4.4.0 or v4.3.6 or even v4.6.0 it doesnt matter) > 3. Now it could be installed by sources following [this guide](https://documentation.wazuh.com/current/deployment-options/wazuh-from-sources/wazuh-server/) or [this file](https://github.com/wazuh/wazuh/blob/11011-prevent-agents-from-lower-version/INSTALL), or a package could be generated with this modified code using the [package generation tool](https://documentation.wazuh.com/current/development/packaging/index.html) (using the flag `--sources <PATH>` indicating the path of this modified source instead of `-b <BRANCH>`) > </details> - <details><summary>Scenario outline 1: Enroll agents to wazuh-manager</summary> ```Gherkin Given a wazuh-manager "v4.5.0". And a <agent-OS> wazuh-agent "<agent-version>". When wazuh-agent tries to enroll to the wazuh-manager. Then the agent "<result>" to wazuh-manager. Examples: | agent-OS | agent-version | result | |----------|---------------|-----------------| | Ubuntu | v4.3.10 | connects | | Ubuntu | v4.4.0 | connects | | Ubuntu | v4.5.0 | connects | | Ubuntu | v4.6.0 | doesn't connect | | Macos | v4.3.10 | connects | | Macos | v4.4.0 | connects | | Macos | v4.5.0 | connects | | Macos | v4.6.0 | doesn't connect | | Windows | v4.3.10 | connects | | Windows | v4.4.0 | connects | | Windows | v4.5.0 | connects | | Windows | v4.6.0 | doesn't connect | ``` </details> - <details><summary>Scenario outline 2: Upgrade manager with agents connected</summary> ```Gherkin Given a wazuh-manager "<manager-initial-version>". And 10 wazuh-agents already connected. When wazuh-manager is upgraded to 4.5.0. Then the upgrade is successful without any error. Examples: | manager-initial-version | |-------------------------| | v4.0.0 | | v4.1.0 | | v4.2.0 | | v4.3.0 | | v4.4.0 | ``` </details> <!-- - <details><summary>Scenario outline 3: Upgrade agent connected to a manager</summary> ```Gherkin Given a wazuh-manager 4.5.0. And a wazuh-agent "<agent-initial-version>" already connected. When wazuh-agent is upgraded to "<agent-final-version>". Then the agent status is "<result>". Examples: | agent-OS | agent-initial-version | agent-final-version | result | |----------|-----------------------|-----------------------|-----------------------| | Ubuntu | v4.4.0 | v4.5.0 | Active | | Ubuntu | v4.5.0 | v4.6.0 | Disconnected | | Macos | v4.4.0 | v4.5.0 | Active | | Macos | v4.5.0 | v4.6.0 | Disconnected | | Windows | v4.4.0 | v4.5.0 | Active | | Windows | v4.5.0 | v4.6.0 | Disconnected | ``` </details> --> ## Expected results <!-- Indicate expected results such as behaviors, logs... --> The manager only accepts connections from agents of equal or lower versions than itself. ## Conclusions and considerations <!-- Indicate considerations to take into account when performing the testing that may not be very intuitive. If any configuration has to be applied, it should be indicated as far as possible. -->
test
validate prevent wazuh agents from connecting to a lower wazuh manager version target version related issue related pr description with this modifications the wazuh manager wont allow an agent with higher version than it to connect preventing compatibility issues although this was not supported before there was no mechanism to avoid it also a new message is logged in agent ossec log file informing this proposed checks note to test some cases with an agent or manager or any version with the new modification the fixed code its needed to create a modified package where the version is manually lowered how to create the modified package clone the branch with the fix git clone b prevent agents from lower version modify the ossec version value in the file src headers defs h line changing it to the desired version it could be or or even it doesnt matter now it could be installed by sources following or or a package could be generated with this modified code using the using the flag sources indicating the path of this modified source instead of b scenario outline enroll agents to wazuh manager gherkin given a wazuh manager and a wazuh agent when wazuh agent tries to enroll to the wazuh manager then the agent to wazuh manager examples agent os agent version result ubuntu connects ubuntu connects ubuntu connects ubuntu doesn t connect macos connects macos connects macos connects macos doesn t connect windows connects windows connects windows connects windows doesn t connect scenario outline upgrade manager with agents connected gherkin given a wazuh manager and wazuh agents already connected when wazuh manager is upgraded to then the upgrade is successful without any error examples manager initial version scenario outline upgrade agent connected to a manager gherkin given a wazuh manager and a wazuh agent already connected when wazuh agent is upgraded to then the agent status is examples agent os agent initial version agent final version result ubuntu active ubuntu disconnected macos active macos disconnected windows active windows disconnected expected results the manager only accepts connections from agents of equal or lower versions than itself conclusions and considerations indicate considerations to take into account when performing the testing that may not be very intuitive if any configuration has to be applied it should be indicated as far as possible
1
90,884
18,270,898,681
IssuesEvent
2021-10-04 13:44:05
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
opened
Provide codefix for MiKo_3054
feature Area: maintainability Area: codefix
Generate a `DependencyProperty` and assign the field's value to it.
1.0
Provide codefix for MiKo_3054 - Generate a `DependencyProperty` and assign the field's value to it.
non_test
provide codefix for miko generate a dependencyproperty and assign the field s value to it
0
95,545
16,096,814,415
IssuesEvent
2021-04-27 01:52:59
eldorplus/portfolio
https://api.github.com/repos/eldorplus/portfolio
opened
CVE-2020-7660 (High) detected in serialize-javascript-1.7.0.tgz
security vulnerability
## CVE-2020-7660 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serialize-javascript-1.7.0.tgz</b></p></summary> <p>Serialize JavaScript to a superset of JSON that includes regular expressions and functions.</p> <p>Library home page: <a href="https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz">https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz</a></p> <p>Path to dependency file: /portfolio/package.json</p> <p>Path to vulnerable library: portfolio/node_modules/serialize-javascript/package.json</p> <p> Dependency Hierarchy: - laravel-mix-0.8.9.tgz (Root Library) - copy-webpack-plugin-4.6.0.tgz - :x: **serialize-javascript-1.7.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> serialize-javascript prior to 3.1.0 allows remote attackers to inject arbitrary code via the function "deleteFunctions" within "index.js". <p>Publish Date: 2020-06-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7660>CVE-2020-7660</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7660">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7660</a></p> <p>Release Date: 2020-06-01</p> <p>Fix Resolution: serialize-javascript - 3.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7660 (High) detected in serialize-javascript-1.7.0.tgz - ## CVE-2020-7660 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serialize-javascript-1.7.0.tgz</b></p></summary> <p>Serialize JavaScript to a superset of JSON that includes regular expressions and functions.</p> <p>Library home page: <a href="https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz">https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz</a></p> <p>Path to dependency file: /portfolio/package.json</p> <p>Path to vulnerable library: portfolio/node_modules/serialize-javascript/package.json</p> <p> Dependency Hierarchy: - laravel-mix-0.8.9.tgz (Root Library) - copy-webpack-plugin-4.6.0.tgz - :x: **serialize-javascript-1.7.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> serialize-javascript prior to 3.1.0 allows remote attackers to inject arbitrary code via the function "deleteFunctions" within "index.js". <p>Publish Date: 2020-06-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7660>CVE-2020-7660</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7660">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7660</a></p> <p>Release Date: 2020-06-01</p> <p>Fix Resolution: serialize-javascript - 3.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in serialize javascript tgz cve high severity vulnerability vulnerable library serialize javascript tgz serialize javascript to a superset of json that includes regular expressions and functions library home page a href path to dependency file portfolio package json path to vulnerable library portfolio node modules serialize javascript package json dependency hierarchy laravel mix tgz root library copy webpack plugin tgz x serialize javascript tgz vulnerable library vulnerability details serialize javascript prior to allows remote attackers to inject arbitrary code via the function deletefunctions within index js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution serialize javascript step up your open source security game with whitesource
0
274,952
23,883,478,958
IssuesEvent
2022-09-08 05:02:32
istio/istio
https://api.github.com/repos/istio/istio
closed
[TF] VM pods do not support host header
area/test and release lifecycle/stale feature/Virtual-machine
### Bug Description This was discovered while working on https://github.com/istio/istio/pull/37914. When specifying an arbitrary host header (e.g. `ok.com`), the VM pod will return a 503. All other pod types return a 200. Here's an example output: ``` I have no name!@a-v1-6994d48868-pk67v:/$ client http://vm.echo1.svc.cluster.local:80 --header="Host:ok.com" [0] Url=http://vm.echo1.svc.cluster.local:80 [0] Latency=1.057692458s [0] ActiveRequests=1 [0] StatusCode=503 [0] ResponseHeader=Content-Length:145 [0] ResponseHeader=Content-Type:text/plain [0] ResponseHeader=Date:Wed, 25 May 2022 19:41:38 GMT [0] ResponseHeader=Server:envoy [0 body] upstream connect error or disconnect/reset before headers. reset reason: connection failure, transport failure reason: delayed connect error: 111 2022-05-25T19:41:39.244342Z info All requests succeeded ``` This request works fine if I leave off the header. It also works fine if I send it to another non-VM pod. ### Version ```prose master ``` ### Additional Information _No response_
1.0
[TF] VM pods do not support host header - ### Bug Description This was discovered while working on https://github.com/istio/istio/pull/37914. When specifying an arbitrary host header (e.g. `ok.com`), the VM pod will return a 503. All other pod types return a 200. Here's an example output: ``` I have no name!@a-v1-6994d48868-pk67v:/$ client http://vm.echo1.svc.cluster.local:80 --header="Host:ok.com" [0] Url=http://vm.echo1.svc.cluster.local:80 [0] Latency=1.057692458s [0] ActiveRequests=1 [0] StatusCode=503 [0] ResponseHeader=Content-Length:145 [0] ResponseHeader=Content-Type:text/plain [0] ResponseHeader=Date:Wed, 25 May 2022 19:41:38 GMT [0] ResponseHeader=Server:envoy [0 body] upstream connect error or disconnect/reset before headers. reset reason: connection failure, transport failure reason: delayed connect error: 111 2022-05-25T19:41:39.244342Z info All requests succeeded ``` This request works fine if I leave off the header. It also works fine if I send it to another non-VM pod. ### Version ```prose master ``` ### Additional Information _No response_
test
vm pods do not support host header bug description this was discovered while working on when specifying an arbitrary host header e g ok com the vm pod will return a all other pod types return a here s an example output i have no name a client header host ok com url latency activerequests statuscode responseheader content length responseheader content type text plain responseheader date wed may gmt responseheader server envoy upstream connect error or disconnect reset before headers reset reason connection failure transport failure reason delayed connect error info all requests succeeded this request works fine if i leave off the header it also works fine if i send it to another non vm pod version prose master additional information no response
1
166,811
12,972,867,978
IssuesEvent
2020-07-21 13:12:55
SenseNet/sn-client
https://api.github.com/repos/SenseNet/sn-client
opened
[E2E test] Drawer items navigation
test
# 🧪E2E test cases The scope of these tests is to ensure that the drawer items link to the desired pages. ![drawer](https://user-images.githubusercontent.com/8918642/88055633-9b890000-cb5f-11ea-91aa-5186d42ad11e.PNG) # 😎 Role All test should run as **admin**. # Test case 1 ## 🧫 Purpose of the test Click on the search (first from top in the picture) icon on the drawer should navigate to the Saved Queries page ## 🐾 Steps 1. Login with admin role 2. Click on the search icon on the drawer ## Expected result - browser navigated to the /saved-queries path # Test case 2 ## 🧫 Purpose of the test Click on the globe (second from top in the picture) icon on the drawer should navigate to the Content page ## 🐾 Steps 1. Login with admin role 2. Click on the search icon on the drawer ## Expected result - browser navigated to the /content/explorer path # Test case 3 ## 🧫 Purpose of the test Click on the people (third from top in the picture) icon on the drawer should navigate to the Users & Groups page ## 🐾 Steps 1. Login with admin role 2. Click on the people icon on the drawer ## Expected result - browser navigated to the /users-and-groups/explorer path # Test case 4 ## 🧫 Purpose of the test Click on the trash (fourth from top in the picture) icon on the drawer should navigate to the Trash page ## 🐾 Steps 1. Login with admin role 2. Click on the trash icon on the drawer ## Expected result - browser navigated to the /trash/explorer path # Test case 5 ## 🧫 Purpose of the test Click on the widgets (fifth from top in the picture) icon on the drawer should navigate to the Content Types page ## 🐾 Steps 1. Login with admin role 2. Click on the widgets icon on the drawer ## Expected result - browser navigated to the /content-types/explorer path # Test case 6 ## 🧫 Purpose of the test Click on the language (sixth from top in the picture) icon on the drawer should navigate to the Content Types page ## 🐾 Steps 1. Login with admin role 2. Click on the language icon on the drawer ## Expected result - browser navigated to the /localization/explorer path # Test case 7 ## 🧫 Purpose of the test Click on the build (seventh from top in the picture) icon on the drawer should navigate to the Setup page ## 🐾 Steps 1. Login with admin role 2. Click on the build icon on the drawer ## Expected result - browser navigated to the /setup path
1.0
[E2E test] Drawer items navigation - # 🧪E2E test cases The scope of these tests is to ensure that the drawer items link to the desired pages. ![drawer](https://user-images.githubusercontent.com/8918642/88055633-9b890000-cb5f-11ea-91aa-5186d42ad11e.PNG) # 😎 Role All test should run as **admin**. # Test case 1 ## 🧫 Purpose of the test Click on the search (first from top in the picture) icon on the drawer should navigate to the Saved Queries page ## 🐾 Steps 1. Login with admin role 2. Click on the search icon on the drawer ## Expected result - browser navigated to the /saved-queries path # Test case 2 ## 🧫 Purpose of the test Click on the globe (second from top in the picture) icon on the drawer should navigate to the Content page ## 🐾 Steps 1. Login with admin role 2. Click on the search icon on the drawer ## Expected result - browser navigated to the /content/explorer path # Test case 3 ## 🧫 Purpose of the test Click on the people (third from top in the picture) icon on the drawer should navigate to the Users & Groups page ## 🐾 Steps 1. Login with admin role 2. Click on the people icon on the drawer ## Expected result - browser navigated to the /users-and-groups/explorer path # Test case 4 ## 🧫 Purpose of the test Click on the trash (fourth from top in the picture) icon on the drawer should navigate to the Trash page ## 🐾 Steps 1. Login with admin role 2. Click on the trash icon on the drawer ## Expected result - browser navigated to the /trash/explorer path # Test case 5 ## 🧫 Purpose of the test Click on the widgets (fifth from top in the picture) icon on the drawer should navigate to the Content Types page ## 🐾 Steps 1. Login with admin role 2. Click on the widgets icon on the drawer ## Expected result - browser navigated to the /content-types/explorer path # Test case 6 ## 🧫 Purpose of the test Click on the language (sixth from top in the picture) icon on the drawer should navigate to the Content Types page ## 🐾 Steps 1. Login with admin role 2. Click on the language icon on the drawer ## Expected result - browser navigated to the /localization/explorer path # Test case 7 ## 🧫 Purpose of the test Click on the build (seventh from top in the picture) icon on the drawer should navigate to the Setup page ## 🐾 Steps 1. Login with admin role 2. Click on the build icon on the drawer ## Expected result - browser navigated to the /setup path
test
drawer items navigation 🧪 test cases the scope of these tests is to ensure that the drawer items link to the desired pages 😎 role all test should run as admin test case 🧫 purpose of the test click on the search first from top in the picture icon on the drawer should navigate to the saved queries page 🐾 steps login with admin role click on the search icon on the drawer expected result browser navigated to the saved queries path test case 🧫 purpose of the test click on the globe second from top in the picture icon on the drawer should navigate to the content page 🐾 steps login with admin role click on the search icon on the drawer expected result browser navigated to the content explorer path test case 🧫 purpose of the test click on the people third from top in the picture icon on the drawer should navigate to the users groups page 🐾 steps login with admin role click on the people icon on the drawer expected result browser navigated to the users and groups explorer path test case 🧫 purpose of the test click on the trash fourth from top in the picture icon on the drawer should navigate to the trash page 🐾 steps login with admin role click on the trash icon on the drawer expected result browser navigated to the trash explorer path test case 🧫 purpose of the test click on the widgets fifth from top in the picture icon on the drawer should navigate to the content types page 🐾 steps login with admin role click on the widgets icon on the drawer expected result browser navigated to the content types explorer path test case 🧫 purpose of the test click on the language sixth from top in the picture icon on the drawer should navigate to the content types page 🐾 steps login with admin role click on the language icon on the drawer expected result browser navigated to the localization explorer path test case 🧫 purpose of the test click on the build seventh from top in the picture icon on the drawer should navigate to the setup page 🐾 steps login with admin role click on the build icon on the drawer expected result browser navigated to the setup path
1
56,023
13,744,656,796
IssuesEvent
2020-10-06 00:35:10
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Docker with GPU 2.3rc0 CUDA runtime implicit initialization on GPU:0 failed. Status: device kernel image is invalid
TF 2.3 subtype: ubuntu/linux type:build/install
It seem that the Docker image tensorflow/tensorflow:2.3.0rc0-gpu won't work with my GPU **BUT** on the other hand the image tensorflow/tensorflow:2.2.0rc0-gpu works fine Or in other words, the solution to the present issue was to "downgrade" to tensorflow/tensorflow:2.2.0rc0-gpu tensorflow/tensorflow:2.3.0rc0-gpu also works fine with CPU only. **System information** - Ubuntu 20.4 - TensorFlow through Docker - TensorFlow version (use command below): - GPU model and memory: Geforce GTX 960M, coreClock: 1.176GHz coreCount: 5 deviceMemorySize: 1.96GiB deviceMemoryBandwidth: 74.65GiB/s - GPU drivers: 440.100 **how to reproduce** ``` > docker run -it --rm --gpus all --entrypoint bash tensorflow/tensorflow:2.3.0rc0-gpu > python >>> import tensorflow as tf >>> inputs = tf.keras.layers.Input(shape=(None,), name="input") >>> embedded = tf.keras.layers.Embedding(100, 16)(inputs) ``` **full stack trace:** ``` 2020-07-06 18:46:55.604377: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcuda.so.1 2020-07-06 18:46:55.608404: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.608911: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1716] Found device 0 with properties: pciBusID: 0000:01:00.0 name: GeForce GTX 960M computeCapability: 5.0 coreClock: 1.176GHz coreCount: 5 deviceMemorySize: 1.96GiB deviceMemoryBandwidth: 74.65GiB/s 2020-07-06 18:46:55.608943: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 2020-07-06 18:46:55.610544: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10 2020-07-06 18:46:55.611696: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10 2020-07-06 18:46:55.611988: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10 2020-07-06 18:46:55.613589: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10 2020-07-06 18:46:55.614478: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10 2020-07-06 18:46:55.618025: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7 2020-07-06 18:46:55.618159: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.618734: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.619206: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1858] Adding visible gpu devices: 0 2020-07-06 18:46:55.619480: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX2 FMA To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags. 2020-07-06 18:46:55.643133: I tensorflow/core/platform/profile_utils/cpu_utils.cc:104] CPU Frequency: 2693910000 Hz 2020-07-06 18:46:55.643781: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x44161a0 initialized for platform Host (this does not guarantee that XLA will be used). Devices: 2020-07-06 18:46:55.643809: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version 2020-07-06 18:46:55.725002: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.725324: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x44aa610 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2020-07-06 18:46:55.725349: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): GeForce GTX 960M, Compute Capability 5.0 2020-07-06 18:46:55.725532: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.725767: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1716] Found device 0 with properties: pciBusID: 0000:01:00.0 name: GeForce GTX 960M computeCapability: 5.0 coreClock: 1.176GHz coreCount: 5 deviceMemorySize: 1.96GiB deviceMemoryBandwidth: 74.65GiB/s 2020-07-06 18:46:55.725796: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 2020-07-06 18:46:55.725828: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10 2020-07-06 18:46:55.725854: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10 2020-07-06 18:46:55.725882: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10 2020-07-06 18:46:55.725908: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10 2020-07-06 18:46:55.725938: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10 2020-07-06 18:46:55.725988: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7 2020-07-06 18:46:55.726091: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.726485: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.726724: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1858] Adding visible gpu devices: 0 2020-07-06 18:46:55.726756: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/engine/base_layer.py", line 926, in __call__ input_list) File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/engine/base_layer.py", line 1098, in _functional_construction_call self._maybe_build(inputs) File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/engine/base_layer.py", line 2643, in _maybe_build self.build(input_shapes) # pylint:disable=not-callable File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/utils/tf_utils.py", line 323, in wrapper output_shape = fn(instance, input_shape) File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/layers/embeddings.py", line 135, in build if (context.executing_eagerly() and context.context().num_gpus() and File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/eager/context.py", line 1082, in num_gpus self.ensure_initialized() File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/eager/context.py", line 539, in ensure_initialized context_handle = pywrap_tfe.TFE_NewContext(opts) tensorflow.python.framework.errors_impl.InternalError: CUDA runtime implicit initialization on GPU:0 failed. Status: device kernel image is invalid ```
1.0
Docker with GPU 2.3rc0 CUDA runtime implicit initialization on GPU:0 failed. Status: device kernel image is invalid - It seem that the Docker image tensorflow/tensorflow:2.3.0rc0-gpu won't work with my GPU **BUT** on the other hand the image tensorflow/tensorflow:2.2.0rc0-gpu works fine Or in other words, the solution to the present issue was to "downgrade" to tensorflow/tensorflow:2.2.0rc0-gpu tensorflow/tensorflow:2.3.0rc0-gpu also works fine with CPU only. **System information** - Ubuntu 20.4 - TensorFlow through Docker - TensorFlow version (use command below): - GPU model and memory: Geforce GTX 960M, coreClock: 1.176GHz coreCount: 5 deviceMemorySize: 1.96GiB deviceMemoryBandwidth: 74.65GiB/s - GPU drivers: 440.100 **how to reproduce** ``` > docker run -it --rm --gpus all --entrypoint bash tensorflow/tensorflow:2.3.0rc0-gpu > python >>> import tensorflow as tf >>> inputs = tf.keras.layers.Input(shape=(None,), name="input") >>> embedded = tf.keras.layers.Embedding(100, 16)(inputs) ``` **full stack trace:** ``` 2020-07-06 18:46:55.604377: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcuda.so.1 2020-07-06 18:46:55.608404: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.608911: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1716] Found device 0 with properties: pciBusID: 0000:01:00.0 name: GeForce GTX 960M computeCapability: 5.0 coreClock: 1.176GHz coreCount: 5 deviceMemorySize: 1.96GiB deviceMemoryBandwidth: 74.65GiB/s 2020-07-06 18:46:55.608943: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 2020-07-06 18:46:55.610544: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10 2020-07-06 18:46:55.611696: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10 2020-07-06 18:46:55.611988: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10 2020-07-06 18:46:55.613589: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10 2020-07-06 18:46:55.614478: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10 2020-07-06 18:46:55.618025: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7 2020-07-06 18:46:55.618159: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.618734: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.619206: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1858] Adding visible gpu devices: 0 2020-07-06 18:46:55.619480: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX2 FMA To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags. 2020-07-06 18:46:55.643133: I tensorflow/core/platform/profile_utils/cpu_utils.cc:104] CPU Frequency: 2693910000 Hz 2020-07-06 18:46:55.643781: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x44161a0 initialized for platform Host (this does not guarantee that XLA will be used). Devices: 2020-07-06 18:46:55.643809: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version 2020-07-06 18:46:55.725002: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.725324: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x44aa610 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2020-07-06 18:46:55.725349: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): GeForce GTX 960M, Compute Capability 5.0 2020-07-06 18:46:55.725532: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.725767: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1716] Found device 0 with properties: pciBusID: 0000:01:00.0 name: GeForce GTX 960M computeCapability: 5.0 coreClock: 1.176GHz coreCount: 5 deviceMemorySize: 1.96GiB deviceMemoryBandwidth: 74.65GiB/s 2020-07-06 18:46:55.725796: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 2020-07-06 18:46:55.725828: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcublas.so.10 2020-07-06 18:46:55.725854: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcufft.so.10 2020-07-06 18:46:55.725882: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcurand.so.10 2020-07-06 18:46:55.725908: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusolver.so.10 2020-07-06 18:46:55.725938: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcusparse.so.10 2020-07-06 18:46:55.725988: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudnn.so.7 2020-07-06 18:46:55.726091: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.726485: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:982] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2020-07-06 18:46:55.726724: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1858] Adding visible gpu devices: 0 2020-07-06 18:46:55.726756: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1 Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/engine/base_layer.py", line 926, in __call__ input_list) File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/engine/base_layer.py", line 1098, in _functional_construction_call self._maybe_build(inputs) File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/engine/base_layer.py", line 2643, in _maybe_build self.build(input_shapes) # pylint:disable=not-callable File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/utils/tf_utils.py", line 323, in wrapper output_shape = fn(instance, input_shape) File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/keras/layers/embeddings.py", line 135, in build if (context.executing_eagerly() and context.context().num_gpus() and File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/eager/context.py", line 1082, in num_gpus self.ensure_initialized() File "/usr/local/lib/python3.6/dist-packages/tensorflow/python/eager/context.py", line 539, in ensure_initialized context_handle = pywrap_tfe.TFE_NewContext(opts) tensorflow.python.framework.errors_impl.InternalError: CUDA runtime implicit initialization on GPU:0 failed. Status: device kernel image is invalid ```
non_test
docker with gpu cuda runtime implicit initialization on gpu failed status device kernel image is invalid it seem that the docker image tensorflow tensorflow gpu won t work with my gpu but on the other hand the image tensorflow tensorflow gpu works fine or in other words the solution to the present issue was to downgrade to tensorflow tensorflow gpu tensorflow tensorflow gpu also works fine with cpu only system information ubuntu tensorflow through docker tensorflow version use command below gpu model and memory geforce gtx coreclock corecount devicememorysize devicememorybandwidth s gpu drivers how to reproduce docker run it rm gpus all entrypoint bash tensorflow tensorflow gpu python import tensorflow as tf inputs tf keras layers input shape none name input embedded tf keras layers embedding inputs full stack trace i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcuda so i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name geforce gtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcublas so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcufft so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcurand so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusolver so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusparse so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudnn so i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow core platform cpu feature guard cc this tensorflow binary is optimized with oneapi deep neural network library onednn to use the following cpu instructions in performance critical operations fma to enable them in other operations rebuild tensorflow with the appropriate compiler flags i tensorflow core platform profile utils cpu utils cc cpu frequency hz i tensorflow compiler xla service service cc xla service initialized for platform host this does not guarantee that xla will be used devices i tensorflow compiler xla service service cc streamexecutor device host default version i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow compiler xla service service cc xla service initialized for platform cuda this does not guarantee that xla will be used devices i tensorflow compiler xla service service cc streamexecutor device geforce gtx compute capability i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name geforce gtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcublas so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcufft so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcurand so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusolver so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusparse so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudnn so i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so traceback most recent call last file line in file usr local lib dist packages tensorflow python keras engine base layer py line in call input list file usr local lib dist packages tensorflow python keras engine base layer py line in functional construction call self maybe build inputs file usr local lib dist packages tensorflow python keras engine base layer py line in maybe build self build input shapes pylint disable not callable file usr local lib dist packages tensorflow python keras utils tf utils py line in wrapper output shape fn instance input shape file usr local lib dist packages tensorflow python keras layers embeddings py line in build if context executing eagerly and context context num gpus and file usr local lib dist packages tensorflow python eager context py line in num gpus self ensure initialized file usr local lib dist packages tensorflow python eager context py line in ensure initialized context handle pywrap tfe tfe newcontext opts tensorflow python framework errors impl internalerror cuda runtime implicit initialization on gpu failed status device kernel image is invalid
0
425,291
12,338,459,792
IssuesEvent
2020-05-14 16:29:15
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
opened
Adjust hearing time intervals to allow 45 minute hearings
Priority: Medium Product: caseflow-hearings Stakeholder: BVA Team: Tango 💃
## User story As a Judge, I need different hearing time increments, so that I have enough time to hold the hearing and have preparation time between hearings. ## Acceptance criteria - [ ] This feature should be accessible to the following user groups: Hearing Coordinators - [ ] Include screenshot(s) in the Github issue - [ ] Update documentation: [link] ## Release notes ### Background/context There have been multiple users requesting 45 minute time slots for scheduling virtual hearings so that VLJs have enough time during and between hearings. Hearing time increments are currently 30 minutes. ### Technical notes Caseflow uses (and cares about starting times) and doesn't track length of time ### Resources/other links <!-- E.g. links to other issues, PRs, Sentry alerts, or Slack threads, or external service requests. -->
1.0
Adjust hearing time intervals to allow 45 minute hearings - ## User story As a Judge, I need different hearing time increments, so that I have enough time to hold the hearing and have preparation time between hearings. ## Acceptance criteria - [ ] This feature should be accessible to the following user groups: Hearing Coordinators - [ ] Include screenshot(s) in the Github issue - [ ] Update documentation: [link] ## Release notes ### Background/context There have been multiple users requesting 45 minute time slots for scheduling virtual hearings so that VLJs have enough time during and between hearings. Hearing time increments are currently 30 minutes. ### Technical notes Caseflow uses (and cares about starting times) and doesn't track length of time ### Resources/other links <!-- E.g. links to other issues, PRs, Sentry alerts, or Slack threads, or external service requests. -->
non_test
adjust hearing time intervals to allow minute hearings user story as a judge i need different hearing time increments so that i have enough time to hold the hearing and have preparation time between hearings acceptance criteria this feature should be accessible to the following user groups hearing coordinators include screenshot s in the github issue update documentation release notes background context there have been multiple users requesting minute time slots for scheduling virtual hearings so that vljs have enough time during and between hearings hearing time increments are currently minutes technical notes caseflow uses and cares about starting times and doesn t track length of time resources other links
0
117,387
9,933,609,102
IssuesEvent
2019-07-02 12:45:25
salesforce/design-system-react
https://api.github.com/repos/salesforce/design-system-react
closed
Remove image snapshot tests from CI
test task
For the amount of time that this takes, the benefit is not enough. These tests are still helpful when updating SLDS versions though.
1.0
Remove image snapshot tests from CI - For the amount of time that this takes, the benefit is not enough. These tests are still helpful when updating SLDS versions though.
test
remove image snapshot tests from ci for the amount of time that this takes the benefit is not enough these tests are still helpful when updating slds versions though
1
310,861
23,358,075,540
IssuesEvent
2022-08-10 09:12:43
web3ui/web3uikit
https://api.github.com/repos/web3ui/web3uikit
opened
[README] updated to for V1
documentation help wanted ready for dev
the read me lost all the content when it became out of date which leads to lots of questions on discord
1.0
[README] updated to for V1 - the read me lost all the content when it became out of date which leads to lots of questions on discord
non_test
updated to for the read me lost all the content when it became out of date which leads to lots of questions on discord
0
247,394
18,857,691,122
IssuesEvent
2021-11-12 08:55:04
luodan01/pe
https://api.github.com/repos/luodan01/pe
opened
Command example for add property does not work as expected
type.DocumentationBug severity.Low
On page 14 of the UG, the following command does not work as expected when copy pasted due to the unnecessary line break ![image.png](https://raw.githubusercontent.com/luodan01/pe/main/files/914bd181-5544-48a4-a821-ceddc6f8ca2e.png) After the user hits enter after copy pasting the command, the following error occurs: ![image.png](https://raw.githubusercontent.com/luodan01/pe/main/files/f64e37a5-14ce-4bf1-9a65-6b40ad274f2a.png) <!--session: 1636704195938-a1e695f1-bbc8-43e4-b787-6fb3c0c40115--> <!--Version: Web v3.4.1-->
1.0
Command example for add property does not work as expected - On page 14 of the UG, the following command does not work as expected when copy pasted due to the unnecessary line break ![image.png](https://raw.githubusercontent.com/luodan01/pe/main/files/914bd181-5544-48a4-a821-ceddc6f8ca2e.png) After the user hits enter after copy pasting the command, the following error occurs: ![image.png](https://raw.githubusercontent.com/luodan01/pe/main/files/f64e37a5-14ce-4bf1-9a65-6b40ad274f2a.png) <!--session: 1636704195938-a1e695f1-bbc8-43e4-b787-6fb3c0c40115--> <!--Version: Web v3.4.1-->
non_test
command example for add property does not work as expected on page of the ug the following command does not work as expected when copy pasted due to the unnecessary line break after the user hits enter after copy pasting the command the following error occurs
0
222,550
17,461,061,040
IssuesEvent
2021-08-06 10:30:21
mswjs/msw
https://api.github.com/repos/mswjs/msw
closed
Unhandled request throw error: [MSW] Failed to mock a "POST" request to
bug good first issue needs:tests scope:browser
## Environment | Name | Version | | ------- | ------- | | msw | 0.33.1 | | browser | Brave Version 1.27.109 Chromium: 92.0.4515.115 (Official Build) (64-bit) | | OS | Linux Mint 20.1 | ## Request handlers <!-- Please provide your mocking setup and the request handlers used --> ### handlers.ts ```js import { rest } from 'msw'; import WebChatConfig from '~/Conversation/webchatConfig/__tests__/fixtures'; export const handlers = [ rest.get( 'https://dev-api-legacy.tolk.ai/v1/webchat/:webchatId', (req, res, ctx) => { console.info(`serving mock for ${req.params.webchatId}`, WebChatConfig); return res(ctx.status(200), ctx.json(WebChatConfig)); } ), ]; ``` ### browser.ts ``` import { setupWorker } from 'msw'; import { handlers } from './handlers'; // This configures a Service Worker with the given request handlers. export const worker = setupWorker(...handlers); ``` ## Actual request I reckon it's from CRA ## Current behavior ``` [MSW] Failed to mock a "POST" request to "https://dev-webchat-analytics.tolk.ai/webchat/analytics/7A3e3aE19b5DF96326DCe6Cf/252eEd2BB6dEEc2d14D39Efe/TKaDsIEm9UXC/uUkFTz7Xr94e0e8UK89nR-a/event/init": TypeError: Failed to fetch POST https://dev-webchat-analytics.tolk.ai/webchat/analytics/7A3e3aE19b5DF96326DCe6Cf/252eEd2BB6dEEc2d14D39Efe/TKaDsIEm9UXC/uUkFTz7Xr94e0e8UK89nR-a/event/init net::ERR_FAILED [WDS] Disconnected! [MSW] Failed to mock a "GET" request to "http://192.168.1.8:3000/sockjs-node/info?t=1627636803942": TypeError: Failed to fetch [MSW] Failed to mock a "GET" request to "http://192.168.1.8:3000/sockjs-node/info?t=1627636814981": TypeError: Failed to fetch [MSW] Failed to mock a "GET" request to "http://192.168.1.8:3000/sockjs-node/info?t=1627636824069": TypeError: Failed to fetch ``` ## Expected behavior I don't want to write handlers for every requests, just the one I want to mock. How do I toggle off error for the unhandled requests ? ## Screenshots <!-- If applicable, attach screenshots to help explain the issue -->
1.0
Unhandled request throw error: [MSW] Failed to mock a "POST" request to - ## Environment | Name | Version | | ------- | ------- | | msw | 0.33.1 | | browser | Brave Version 1.27.109 Chromium: 92.0.4515.115 (Official Build) (64-bit) | | OS | Linux Mint 20.1 | ## Request handlers <!-- Please provide your mocking setup and the request handlers used --> ### handlers.ts ```js import { rest } from 'msw'; import WebChatConfig from '~/Conversation/webchatConfig/__tests__/fixtures'; export const handlers = [ rest.get( 'https://dev-api-legacy.tolk.ai/v1/webchat/:webchatId', (req, res, ctx) => { console.info(`serving mock for ${req.params.webchatId}`, WebChatConfig); return res(ctx.status(200), ctx.json(WebChatConfig)); } ), ]; ``` ### browser.ts ``` import { setupWorker } from 'msw'; import { handlers } from './handlers'; // This configures a Service Worker with the given request handlers. export const worker = setupWorker(...handlers); ``` ## Actual request I reckon it's from CRA ## Current behavior ``` [MSW] Failed to mock a "POST" request to "https://dev-webchat-analytics.tolk.ai/webchat/analytics/7A3e3aE19b5DF96326DCe6Cf/252eEd2BB6dEEc2d14D39Efe/TKaDsIEm9UXC/uUkFTz7Xr94e0e8UK89nR-a/event/init": TypeError: Failed to fetch POST https://dev-webchat-analytics.tolk.ai/webchat/analytics/7A3e3aE19b5DF96326DCe6Cf/252eEd2BB6dEEc2d14D39Efe/TKaDsIEm9UXC/uUkFTz7Xr94e0e8UK89nR-a/event/init net::ERR_FAILED [WDS] Disconnected! [MSW] Failed to mock a "GET" request to "http://192.168.1.8:3000/sockjs-node/info?t=1627636803942": TypeError: Failed to fetch [MSW] Failed to mock a "GET" request to "http://192.168.1.8:3000/sockjs-node/info?t=1627636814981": TypeError: Failed to fetch [MSW] Failed to mock a "GET" request to "http://192.168.1.8:3000/sockjs-node/info?t=1627636824069": TypeError: Failed to fetch ``` ## Expected behavior I don't want to write handlers for every requests, just the one I want to mock. How do I toggle off error for the unhandled requests ? ## Screenshots <!-- If applicable, attach screenshots to help explain the issue -->
test
unhandled request throw error failed to mock a post request to environment name version msw browser brave version chromium official build bit os linux mint request handlers handlers ts js import rest from msw import webchatconfig from conversation webchatconfig tests fixtures export const handlers rest get req res ctx console info serving mock for req params webchatid webchatconfig return res ctx status ctx json webchatconfig browser ts import setupworker from msw import handlers from handlers this configures a service worker with the given request handlers export const worker setupworker handlers actual request i reckon it s from cra current behavior failed to mock a post request to typeerror failed to fetch post net err failed disconnected failed to mock a get request to typeerror failed to fetch failed to mock a get request to typeerror failed to fetch failed to mock a get request to typeerror failed to fetch expected behavior i don t want to write handlers for every requests just the one i want to mock how do i toggle off error for the unhandled requests screenshots
1
219,691
16,848,916,366
IssuesEvent
2021-06-20 04:42:42
mrmcc3/my-blog
https://api.github.com/repos/mrmcc3/my-blog
closed
Fix Readme
documentation
fix the readme. cleanup the repo a bit with the idea that this could be public. especially after #6
1.0
Fix Readme - fix the readme. cleanup the repo a bit with the idea that this could be public. especially after #6
non_test
fix readme fix the readme cleanup the repo a bit with the idea that this could be public especially after
0
228,994
18,275,281,490
IssuesEvent
2021-10-04 18:03:37
ToriLemons/QA-Practice-One
https://api.github.com/repos/ToriLemons/QA-Practice-One
opened
Acceptance Criteria-Website 1-JH
acceptance test
PM Acceptance Criteria 1. There was a header and footer • PASSED • Notes: 2. There was menu access • PASSED • Notes: 3. There was at least 4 social media icons • PASSED • Notes: NO. Only 3 listed 4. There was a logo • PASSED • Notes: 5. There was a consistant theme • PASSED • Notes: 6. The color scheme was ADA compliant • PASSED • Notes: 83% LH Report. 3 Contrast Errors on WAVE, 3 Errors on Code Sniffer for WCAG 2.1 AA 7. There was a contact page with a form • PASSED • Notes: 8. There was a google map embed on the contact page • PASSED • Notes: 9. There were at least 5 images with a border • PASSED • Notes: 10. All of the links on the website function as intended. • - Check if passed. • Notes: NO. Pet Name Lemon Image came up with Site can’t be reached ERROR Customer Acceptance 1. Tab labels are correct • PASSED • Notes: 2. Current page is clear on nav bar • PASSED • Notes: 3. Spelling is correct • PASSED • Notes: No. 1 item in Contact Pull down is not only duplicate, but also misspelled. MOR as opposed to MORE. 4. Alt text is appropriate and makes sense • PASSED • Notes: Not all links are identifiable . Social Media Icons, Full name, and Email are not identified when using a Screen Reader. 5. While it isnt made for mobile, it does not break on resizing and no horizonal scrolling ever • PASSED • Notes: 6. Tabbing works and is visible • PASSED • Notes: 7. Enlarging text does not break format • PASSED • Notes: 8. Consistant navigation • PASSED • Notes:
1.0
Acceptance Criteria-Website 1-JH - PM Acceptance Criteria 1. There was a header and footer • PASSED • Notes: 2. There was menu access • PASSED • Notes: 3. There was at least 4 social media icons • PASSED • Notes: NO. Only 3 listed 4. There was a logo • PASSED • Notes: 5. There was a consistant theme • PASSED • Notes: 6. The color scheme was ADA compliant • PASSED • Notes: 83% LH Report. 3 Contrast Errors on WAVE, 3 Errors on Code Sniffer for WCAG 2.1 AA 7. There was a contact page with a form • PASSED • Notes: 8. There was a google map embed on the contact page • PASSED • Notes: 9. There were at least 5 images with a border • PASSED • Notes: 10. All of the links on the website function as intended. • - Check if passed. • Notes: NO. Pet Name Lemon Image came up with Site can’t be reached ERROR Customer Acceptance 1. Tab labels are correct • PASSED • Notes: 2. Current page is clear on nav bar • PASSED • Notes: 3. Spelling is correct • PASSED • Notes: No. 1 item in Contact Pull down is not only duplicate, but also misspelled. MOR as opposed to MORE. 4. Alt text is appropriate and makes sense • PASSED • Notes: Not all links are identifiable . Social Media Icons, Full name, and Email are not identified when using a Screen Reader. 5. While it isnt made for mobile, it does not break on resizing and no horizonal scrolling ever • PASSED • Notes: 6. Tabbing works and is visible • PASSED • Notes: 7. Enlarging text does not break format • PASSED • Notes: 8. Consistant navigation • PASSED • Notes:
test
acceptance criteria website jh pm acceptance criteria there was a header and footer • passed • notes there was menu access • passed • notes there was at least social media icons • passed • notes no only listed there was a logo • passed • notes there was a consistant theme • passed • notes the color scheme was ada compliant • passed • notes lh report contrast errors on wave errors on code sniffer for wcag aa there was a contact page with a form • passed • notes there was a google map embed on the contact page • passed • notes there were at least images with a border • passed • notes all of the links on the website function as intended • check if passed • notes no pet name lemon image came up with site can’t be reached error customer acceptance tab labels are correct • passed • notes current page is clear on nav bar • passed • notes spelling is correct • passed • notes no item in contact pull down is not only duplicate but also misspelled mor as opposed to more alt text is appropriate and makes sense • passed • notes not all links are identifiable social media icons full name and email are not identified when using a screen reader while it isnt made for mobile it does not break on resizing and no horizonal scrolling ever • passed • notes tabbing works and is visible • passed • notes enlarging text does not break format • passed • notes consistant navigation • passed • notes
1
828,333
31,822,250,301
IssuesEvent
2023-09-14 03:59:42
calcom/cal.com
https://api.github.com/repos/calcom/cal.com
closed
Email not dispatched on successful payment
🐛 bug High priority billing emails
### Issue Summary A summary of the issue. This needs to be a clear detailed-rich summary. ### Steps to Reproduce 1. Install stripe app and connect user. 2. Enable stripe in event-type 3. Create a booking for that event-type. [Video](https://youtu.be/i_C_CYdIa2Q)
1.0
Email not dispatched on successful payment - ### Issue Summary A summary of the issue. This needs to be a clear detailed-rich summary. ### Steps to Reproduce 1. Install stripe app and connect user. 2. Enable stripe in event-type 3. Create a booking for that event-type. [Video](https://youtu.be/i_C_CYdIa2Q)
non_test
email not dispatched on successful payment issue summary a summary of the issue this needs to be a clear detailed rich summary steps to reproduce install stripe app and connect user enable stripe in event type create a booking for that event type
0
100,103
4,076,271,736
IssuesEvent
2016-05-29 20:01:06
FormalHellhound/BruhBot
https://api.github.com/repos/FormalHellhound/BruhBot
closed
Fancy Tag Messages
Implemented Feature Low Priority Feature Request
Inspired by Onee-chan. Onee-chan has two features that drew my eye - `-letsplay` and `-call`. Both messages tag @everyone and announce that the user who ran the command would like to (in the former) play a game or (in the latter) have a voice chat, accompanied by emoji. I think this sort of prettied up usage-specific tagging method would be fun. Bonus points if these can be configured by the user in the config file to create their own shoutout commands - ex. maybe I run a server where League of Legends is popular and want the command to specifically call out a role of users who also play. In this case the command could be configured to change the role tagged (instead of @everyone) and the flavor text of the message ala `Attention @league , @username would like to play a round of League!!`.
1.0
Fancy Tag Messages - Inspired by Onee-chan. Onee-chan has two features that drew my eye - `-letsplay` and `-call`. Both messages tag @everyone and announce that the user who ran the command would like to (in the former) play a game or (in the latter) have a voice chat, accompanied by emoji. I think this sort of prettied up usage-specific tagging method would be fun. Bonus points if these can be configured by the user in the config file to create their own shoutout commands - ex. maybe I run a server where League of Legends is popular and want the command to specifically call out a role of users who also play. In this case the command could be configured to change the role tagged (instead of @everyone) and the flavor text of the message ala `Attention @league , @username would like to play a round of League!!`.
non_test
fancy tag messages inspired by onee chan onee chan has two features that drew my eye letsplay and call both messages tag everyone and announce that the user who ran the command would like to in the former play a game or in the latter have a voice chat accompanied by emoji i think this sort of prettied up usage specific tagging method would be fun bonus points if these can be configured by the user in the config file to create their own shoutout commands ex maybe i run a server where league of legends is popular and want the command to specifically call out a role of users who also play in this case the command could be configured to change the role tagged instead of everyone and the flavor text of the message ala attention league username would like to play a round of league
0
301,327
26,035,649,747
IssuesEvent
2022-12-22 04:30:26
codestates-seb/seb40_main_001
https://api.github.com/repos/codestates-seb/seb40_main_001
opened
✅ Test: 대댓글 테스트
Back Test
## Description 대댓글 도메인의 테스트 코드를 작성합니다. ## Todo - [ ] Reply 단위 테스트 - [ ] Reply 통합 테스트 ## ETC - 주의 사항 혹은 기타 내용
1.0
✅ Test: 대댓글 테스트 - ## Description 대댓글 도메인의 테스트 코드를 작성합니다. ## Todo - [ ] Reply 단위 테스트 - [ ] Reply 통합 테스트 ## ETC - 주의 사항 혹은 기타 내용
test
✅ test 대댓글 테스트 description 대댓글 도메인의 테스트 코드를 작성합니다 todo reply 단위 테스트 reply 통합 테스트 etc 주의 사항 혹은 기타 내용
1
52,090
6,218,503,674
IssuesEvent
2017-07-09 02:36:01
shader-slang/slang
https://api.github.com/repos/shader-slang/slang
closed
GLSL layout rules for `uniform` and `buffer` blocks
glsl testing
There is code in `TypeLayout.cpp` that tries to implement the `std140` and `std430` rules, but I have little confidence that it is being invoked correctly. Tasks: - Set up some reflection-generation tests for GLSL, so we can see how layouts are being computed - Need to be careful when defining the expected output here; should probably run the same input through `glslang` when generating baselines to double-check offsets - Ensure that we are picking up the rules specified as a `layout` attribute and applying them correctly. - Ensure that given GLSL source we are picking appropriate rules by default when nothing is specified (e.g., `std140` for all `uniform` blocks, and `std430` for all `buffer` blocks) - Make sure to emit downstream code that reflects the layout choices we make, either by applying a `layout` attribute to the block, or by applying `layout(offset=...)` to each member. We should be conservative and try not to require too many extended features that could make it harder to output portable OpenGL GLSL later.
1.0
GLSL layout rules for `uniform` and `buffer` blocks - There is code in `TypeLayout.cpp` that tries to implement the `std140` and `std430` rules, but I have little confidence that it is being invoked correctly. Tasks: - Set up some reflection-generation tests for GLSL, so we can see how layouts are being computed - Need to be careful when defining the expected output here; should probably run the same input through `glslang` when generating baselines to double-check offsets - Ensure that we are picking up the rules specified as a `layout` attribute and applying them correctly. - Ensure that given GLSL source we are picking appropriate rules by default when nothing is specified (e.g., `std140` for all `uniform` blocks, and `std430` for all `buffer` blocks) - Make sure to emit downstream code that reflects the layout choices we make, either by applying a `layout` attribute to the block, or by applying `layout(offset=...)` to each member. We should be conservative and try not to require too many extended features that could make it harder to output portable OpenGL GLSL later.
test
glsl layout rules for uniform and buffer blocks there is code in typelayout cpp that tries to implement the and rules but i have little confidence that it is being invoked correctly tasks set up some reflection generation tests for glsl so we can see how layouts are being computed need to be careful when defining the expected output here should probably run the same input through glslang when generating baselines to double check offsets ensure that we are picking up the rules specified as a layout attribute and applying them correctly ensure that given glsl source we are picking appropriate rules by default when nothing is specified e g for all uniform blocks and for all buffer blocks make sure to emit downstream code that reflects the layout choices we make either by applying a layout attribute to the block or by applying layout offset to each member we should be conservative and try not to require too many extended features that could make it harder to output portable opengl glsl later
1
186,103
14,394,638,184
IssuesEvent
2020-12-03 01:46:13
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
tengteng/Guava: _vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go; 3 LoC
fresh test tiny
Found a possible issue in [tengteng/Guava](https://www.github.com/tengteng/Guava) at [_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 106 may start a goroutine [Click here to see the code in its original context.](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range importerTests { runImporterTest(t, imp, initmap, &test) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: f44ca584d2dc0fe32182990065cfc4fd0e6cebe8
1.0
tengteng/Guava: _vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go; 3 LoC - Found a possible issue in [tengteng/Guava](https://www.github.com/tengteng/Guava) at [_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 106 may start a goroutine [Click here to see the code in its original context.](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range importerTests { runImporterTest(t, imp, initmap, &test) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: f44ca584d2dc0fe32182990065cfc4fd0e6cebe8
test
tengteng guava vendor src golang org x tools go gccgoimporter importer test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test range importertests runimportertest t imp initmap test leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
71,249
18,545,302,453
IssuesEvent
2021-10-21 21:13:41
OpenVisualCloud/SVT-VP9
https://api.github.com/repos/OpenVisualCloud/SVT-VP9
closed
Build error MacOS 10.15.2
build
After building SVT-AV1 and SVT-HEVC, now trying to build on MacOS Catalina 10.15.2 but this one fails with the following error: $ cmake -DCMAKE_INSTALL_PREFIX=/usr/local/bin -DBUILD_SHARED_LIBS=0 ../ -- Found YASM: /usr/local/bin/bin/yasm -- CMAKE_BUILD_TYPE = Release -- CMAKE_C_COMPILER = /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/cc -- CMAKE_ASM_NASM_COMPILER = /usr/local/bin/bin/yasm -- CMAKE_AR = /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ar -- CMAKE_RANLIB = /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib -- CMAKE_INSTALL_PREFIX = /usr/local/bin -- Configuring done -- Generating done -- Build files have been written to: svt-vp9/SVT-VP9/Build $ make [ 30%] Built target libVPX [ 39%] Built target VP9_ASM_AVX2 [ 43%] Built target VP9_C_DEFAULT [ 54%] Built target VP9_ASM_SSE2 [ 58%] Built target VP9_ASM_SSSE3 [ 59%] Built target VP9_ASM_SSE4_1 [ 60%] Building C object Source/Lib/Codec/CMakeFiles/SvtVp9Enc.dir/EbEncHandle.c.o svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:390:39: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? memset(eb_vp9_lp_group, 0, sizeof(eb_vp9_lp_group)); ^~~~~~~~~~~~~~~ eb_vp9_num_groups /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_string.h:77:33: note: expanded from macro 'memset' __builtin___memset_chk (dest, __VA_ARGS__, __darwin_obsz0 (dest)) ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:390:12: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? memset(eb_vp9_lp_group, 0, sizeof(eb_vp9_lp_group)); ^~~~~~~~~~~~~~~ eb_vp9_num_groups /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_string.h:77:62: note: expanded from macro 'memset' __builtin___memset_chk (dest, __VA_ARGS__, __darwin_obsz0 (dest)) ^ /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_common.h:38:55: note: expanded from macro '__darwin_obsz0' #define __darwin_obsz0(object) __builtin_object_size (object, 0) ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:390:12: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? memset(eb_vp9_lp_group, 0, sizeof(eb_vp9_lp_group)); ^~~~~~~~~~~~~~~ eb_vp9_num_groups /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_string.h:77:27: note: expanded from macro 'memset' __builtin___memset_chk (dest, __VA_ARGS__, __darwin_obsz0 (dest)) ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:412:13: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? eb_vp9_lp_group[socket_id].group[eb_vp9_lp_group[socket_id].num++] = processor_id; ^~~~~~~~~~~~~~~ eb_vp9_num_groups svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:412:28: error: subscripted value is not an array, pointer, or vector eb_vp9_lp_group[socket_id].group[eb_vp9_lp_group[socket_id].num++] = processor_id; ~~~~~~~~~~~~~~~^~~~~~~~~~ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:412:46: error: use of undeclared identifier 'eb_vp9_lp_group' eb_vp9_lp_group[socket_id].group[eb_vp9_lp_group[socket_id].num++] = processor_id; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:602:14: warning: unused variable 'num_logical_processors' [-Wunused-variable] uint32_t num_logical_processors = get_num_cores(); ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:600:69: warning: unused parameter 'config_ptr' [-Wunused-parameter] void eb_set_thread_management_parameters( EbSvtVp9EncConfiguration *config_ptr){ ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:2981:17: warning: unused parameter 'msg' [-Wunused-parameter] const char *msg, ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:2983:11: warning: unused parameter 'error' [-Wunused-parameter] Errno error) ^ 4 warnings and 6 errors generated. make[2]: *** [Source/Lib/Codec/CMakeFiles/SvtVp9Enc.dir/EbEncHandle.c.o] Error 1 make[1]: *** [Source/Lib/Codec/CMakeFiles/SvtVp9Enc.dir/all] Error 2 make: *** [all] Error 2 Don't find anything on the net, so I'm asking here and hoping for some help.
1.0
Build error MacOS 10.15.2 - After building SVT-AV1 and SVT-HEVC, now trying to build on MacOS Catalina 10.15.2 but this one fails with the following error: $ cmake -DCMAKE_INSTALL_PREFIX=/usr/local/bin -DBUILD_SHARED_LIBS=0 ../ -- Found YASM: /usr/local/bin/bin/yasm -- CMAKE_BUILD_TYPE = Release -- CMAKE_C_COMPILER = /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/cc -- CMAKE_ASM_NASM_COMPILER = /usr/local/bin/bin/yasm -- CMAKE_AR = /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ar -- CMAKE_RANLIB = /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib -- CMAKE_INSTALL_PREFIX = /usr/local/bin -- Configuring done -- Generating done -- Build files have been written to: svt-vp9/SVT-VP9/Build $ make [ 30%] Built target libVPX [ 39%] Built target VP9_ASM_AVX2 [ 43%] Built target VP9_C_DEFAULT [ 54%] Built target VP9_ASM_SSE2 [ 58%] Built target VP9_ASM_SSSE3 [ 59%] Built target VP9_ASM_SSE4_1 [ 60%] Building C object Source/Lib/Codec/CMakeFiles/SvtVp9Enc.dir/EbEncHandle.c.o svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:390:39: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? memset(eb_vp9_lp_group, 0, sizeof(eb_vp9_lp_group)); ^~~~~~~~~~~~~~~ eb_vp9_num_groups /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_string.h:77:33: note: expanded from macro 'memset' __builtin___memset_chk (dest, __VA_ARGS__, __darwin_obsz0 (dest)) ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:390:12: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? memset(eb_vp9_lp_group, 0, sizeof(eb_vp9_lp_group)); ^~~~~~~~~~~~~~~ eb_vp9_num_groups /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_string.h:77:62: note: expanded from macro 'memset' __builtin___memset_chk (dest, __VA_ARGS__, __darwin_obsz0 (dest)) ^ /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_common.h:38:55: note: expanded from macro '__darwin_obsz0' #define __darwin_obsz0(object) __builtin_object_size (object, 0) ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:390:12: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? memset(eb_vp9_lp_group, 0, sizeof(eb_vp9_lp_group)); ^~~~~~~~~~~~~~~ eb_vp9_num_groups /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/secure/_string.h:77:27: note: expanded from macro 'memset' __builtin___memset_chk (dest, __VA_ARGS__, __darwin_obsz0 (dest)) ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:412:13: error: use of undeclared identifier 'eb_vp9_lp_group'; did you mean 'eb_vp9_num_groups'? eb_vp9_lp_group[socket_id].group[eb_vp9_lp_group[socket_id].num++] = processor_id; ^~~~~~~~~~~~~~~ eb_vp9_num_groups svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:127:34: note: 'eb_vp9_num_groups' declared here uint8_t eb_vp9_num_groups = 0; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:412:28: error: subscripted value is not an array, pointer, or vector eb_vp9_lp_group[socket_id].group[eb_vp9_lp_group[socket_id].num++] = processor_id; ~~~~~~~~~~~~~~~^~~~~~~~~~ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:412:46: error: use of undeclared identifier 'eb_vp9_lp_group' eb_vp9_lp_group[socket_id].group[eb_vp9_lp_group[socket_id].num++] = processor_id; ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:602:14: warning: unused variable 'num_logical_processors' [-Wunused-variable] uint32_t num_logical_processors = get_num_cores(); ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:600:69: warning: unused parameter 'config_ptr' [-Wunused-parameter] void eb_set_thread_management_parameters( EbSvtVp9EncConfiguration *config_ptr){ ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:2981:17: warning: unused parameter 'msg' [-Wunused-parameter] const char *msg, ^ svt-vp9/SVT-VP9/Source/Lib/Codec/EbEncHandle.c:2983:11: warning: unused parameter 'error' [-Wunused-parameter] Errno error) ^ 4 warnings and 6 errors generated. make[2]: *** [Source/Lib/Codec/CMakeFiles/SvtVp9Enc.dir/EbEncHandle.c.o] Error 1 make[1]: *** [Source/Lib/Codec/CMakeFiles/SvtVp9Enc.dir/all] Error 2 make: *** [all] Error 2 Don't find anything on the net, so I'm asking here and hoping for some help.
non_test
build error macos after building svt and svt hevc now trying to build on macos catalina but this one fails with the following error cmake dcmake install prefix usr local bin dbuild shared libs found yasm usr local bin bin yasm cmake build type release cmake c compiler applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin cc cmake asm nasm compiler usr local bin bin yasm cmake ar applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin ar cmake ranlib applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin ranlib cmake install prefix usr local bin configuring done generating done build files have been written to svt svt build make built target libvpx built target asm built target c default built target asm built target asm built target asm building c object source lib codec cmakefiles dir ebenchandle c o svt svt source lib codec ebenchandle c error use of undeclared identifier eb lp group did you mean eb num groups memset eb lp group sizeof eb lp group eb num groups applications xcode app contents developer platforms macosx platform developer sdks sdk usr include secure string h note expanded from macro memset builtin memset chk dest va args darwin dest svt svt source lib codec ebenchandle c note eb num groups declared here t eb num groups svt svt source lib codec ebenchandle c error use of undeclared identifier eb lp group did you mean eb num groups memset eb lp group sizeof eb lp group eb num groups applications xcode app contents developer platforms macosx platform developer sdks sdk usr include secure string h note expanded from macro memset builtin memset chk dest va args darwin dest applications xcode app contents developer platforms macosx platform developer sdks sdk usr include secure common h note expanded from macro darwin define darwin object builtin object size object svt svt source lib codec ebenchandle c note eb num groups declared here t eb num groups svt svt source lib codec ebenchandle c error use of undeclared identifier eb lp group did you mean eb num groups memset eb lp group sizeof eb lp group eb num groups applications xcode app contents developer platforms macosx platform developer sdks sdk usr include secure string h note expanded from macro memset builtin memset chk dest va args darwin dest svt svt source lib codec ebenchandle c note eb num groups declared here t eb num groups svt svt source lib codec ebenchandle c error use of undeclared identifier eb lp group did you mean eb num groups eb lp group group num processor id eb num groups svt svt source lib codec ebenchandle c note eb num groups declared here t eb num groups svt svt source lib codec ebenchandle c error subscripted value is not an array pointer or vector eb lp group group num processor id svt svt source lib codec ebenchandle c error use of undeclared identifier eb lp group eb lp group group num processor id svt svt source lib codec ebenchandle c warning unused variable num logical processors t num logical processors get num cores svt svt source lib codec ebenchandle c warning unused parameter config ptr void eb set thread management parameters config ptr svt svt source lib codec ebenchandle c warning unused parameter msg const char msg svt svt source lib codec ebenchandle c warning unused parameter error errno error warnings and errors generated make error make error make error don t find anything on the net so i m asking here and hoping for some help
0
155,241
12,244,344,976
IssuesEvent
2020-05-05 10:56:57
WoWManiaUK/Redemption
https://api.github.com/repos/WoWManiaUK/Redemption
closed
[Spell/Druid] Frenzied Regeneration must crit
Fix - Ready to Test
Checked by logs, Frenzied Regeneration must have crit chance. In current TC state, it dont have. TC issue about it: https://github.com/TrinityCore/TrinityCore/issues/20798
1.0
[Spell/Druid] Frenzied Regeneration must crit - Checked by logs, Frenzied Regeneration must have crit chance. In current TC state, it dont have. TC issue about it: https://github.com/TrinityCore/TrinityCore/issues/20798
test
frenzied regeneration must crit checked by logs frenzied regeneration must have crit chance in current tc state it dont have tc issue about it
1