Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
267,820 | 23,322,171,776 | IssuesEvent | 2022-08-08 17:26:37 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | closed | Add a test around removing a shipping zone region | package: @woocommerce/e2e-core-tests | We should add a test around removing a shipping zone regions. See: https://github.com/woocommerce/woocommerce/issues/32016 | 1.0 | Add a test around removing a shipping zone region - We should add a test around removing a shipping zone regions. See: https://github.com/woocommerce/woocommerce/issues/32016 | test | add a test around removing a shipping zone region we should add a test around removing a shipping zone regions see | 1 |
222,143 | 17,036,440,504 | IssuesEvent | 2021-07-05 07:48:43 | Decathlon/vitamin-android | https://api.github.com/repos/Decathlon/vitamin-android | closed | Checkboxes module documentation | documentation 📚 | Create a README.md file in the root folder of checkboxes module with:
- for each variant, an usage section with kotlin class, attribute style and the style hardcoded
- standalone instructions | 1.0 | Checkboxes module documentation - Create a README.md file in the root folder of checkboxes module with:
- for each variant, an usage section with kotlin class, attribute style and the style hardcoded
- standalone instructions | non_test | checkboxes module documentation create a readme md file in the root folder of checkboxes module with for each variant an usage section with kotlin class attribute style and the style hardcoded standalone instructions | 0 |
379,382 | 11,220,883,607 | IssuesEvent | 2020-01-07 16:39:45 | idaholab/raven | https://api.github.com/repos/idaholab/raven | closed | [TASK] Add an error if plugins not installed and tests are requested | priority_minor task | --------
Issue Description
--------
**Is your feature request related to a problem? Please describe.**
Add an error if plugins not installed and tests are requested.
**Describe the solution you'd like**
Error when the tests are run with ./run_tests --plugins but none of the plugins are installed
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or task?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| 1.0 | [TASK] Add an error if plugins not installed and tests are requested - --------
Issue Description
--------
**Is your feature request related to a problem? Please describe.**
Add an error if plugins not installed and tests are requested.
**Describe the solution you'd like**
Error when the tests are run with ./run_tests --plugins but none of the plugins are installed
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or task?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| non_test | add an error if plugins not installed and tests are requested issue description is your feature request related to a problem please describe add an error if plugins not installed and tests are requested describe the solution you d like error when the tests are run with run tests plugins but none of the plugins are installed describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided | 0 |
223,485 | 17,602,815,855 | IssuesEvent | 2021-08-17 13:48:48 | cseelhoff/RimThreaded | https://api.github.com/repos/cseelhoff/RimThreaded | closed | Doors Expanded (Dev) red errors ( RT 2.3.6.7 ) | Bug Reproducible Accepted For Testing Mod Incompatibility Confirmed 2.2.X.X 2.3.X.X |
When using a door that isn't a 1x1 for Doors expanded it throws out a red error when being used
**Steps to reproduce the behavior**
1. Install Doors expanded
2. Place a door that isn't a 1x1
3. Draft a pawn or let he pawn use it
4. See error
Screenshots of error happening

**Error Log**
https://gist.github.com/3225bf5e7a43ee6bf7b29be6f5bc1a00
**Mod list (Preferably a RimPy compatible list.)**
* Harmony
* Core
* Royalty (optional)
* HugsLib
* Doors Expanded (Dev)
* RimThreaded ( 2.3.6.7 )
| 1.0 | Doors Expanded (Dev) red errors ( RT 2.3.6.7 ) -
When using a door that isn't a 1x1 for Doors expanded it throws out a red error when being used
**Steps to reproduce the behavior**
1. Install Doors expanded
2. Place a door that isn't a 1x1
3. Draft a pawn or let he pawn use it
4. See error
Screenshots of error happening

**Error Log**
https://gist.github.com/3225bf5e7a43ee6bf7b29be6f5bc1a00
**Mod list (Preferably a RimPy compatible list.)**
* Harmony
* Core
* Royalty (optional)
* HugsLib
* Doors Expanded (Dev)
* RimThreaded ( 2.3.6.7 )
| test | doors expanded dev red errors rt when using a door that isn t a for doors expanded it throws out a red error when being used steps to reproduce the behavior install doors expanded place a door that isn t a draft a pawn or let he pawn use it see error screenshots of error happening error log mod list preferably a rimpy compatible list harmony core royalty optional hugslib doors expanded dev rimthreaded | 1 |
2,889 | 3,228,070,623 | IssuesEvent | 2015-10-11 19:32:26 | bit-team/backintime | https://api.github.com/repos/bit-team/backintime | closed | Ctrl+Q, Ctrl+W should work | Fix Released Low usability | It should be possible to exit the application using ctrl+Q.
------------------------------------
Imported from Launchpad using lp2gh.
* date created: 2010-12-07T18:34:01Z
* owner: kiddo
* the launchpad url was https://bugs.launchpad.net/bugs/686694
| True | Ctrl+Q, Ctrl+W should work - It should be possible to exit the application using ctrl+Q.
------------------------------------
Imported from Launchpad using lp2gh.
* date created: 2010-12-07T18:34:01Z
* owner: kiddo
* the launchpad url was https://bugs.launchpad.net/bugs/686694
| non_test | ctrl q ctrl w should work it should be possible to exit the application using ctrl q imported from launchpad using date created owner kiddo the launchpad url was | 0 |
245,341 | 20,763,426,335 | IssuesEvent | 2022-03-15 18:15:28 | caelum/alura-funcionalidades | https://api.github.com/repos/caelum/alura-funcionalidades | closed | Prova no final de cada Curso, Formação ou Capítulo. | teste | Acredito que uma pequena prova no final de cada capítulo, curso e formação, possa ajudar as pessoas tanto a fixar melhor o conteúdo que viu, assim como aumenta a pratica no conteúdo que o aluno está querendo aprender, dando mais confiança no conteúdo e domínio da linguagem.
um exemplo que posso incluir na ideia: eu mesmo fiz os 4 cursos de html e css, fiz todos os exercicios e no fim resolvi criar um site proprio partindo do codigo que ja tinha e me deparei com algumas situações ou que não lembrava, ou que não que não tinha na aula.
| 1.0 | Prova no final de cada Curso, Formação ou Capítulo. - Acredito que uma pequena prova no final de cada capítulo, curso e formação, possa ajudar as pessoas tanto a fixar melhor o conteúdo que viu, assim como aumenta a pratica no conteúdo que o aluno está querendo aprender, dando mais confiança no conteúdo e domínio da linguagem.
um exemplo que posso incluir na ideia: eu mesmo fiz os 4 cursos de html e css, fiz todos os exercicios e no fim resolvi criar um site proprio partindo do codigo que ja tinha e me deparei com algumas situações ou que não lembrava, ou que não que não tinha na aula.
| test | prova no final de cada curso formação ou capítulo acredito que uma pequena prova no final de cada capítulo curso e formação possa ajudar as pessoas tanto a fixar melhor o conteúdo que viu assim como aumenta a pratica no conteúdo que o aluno está querendo aprender dando mais confiança no conteúdo e domínio da linguagem um exemplo que posso incluir na ideia eu mesmo fiz os cursos de html e css fiz todos os exercicios e no fim resolvi criar um site proprio partindo do codigo que ja tinha e me deparei com algumas situações ou que não lembrava ou que não que não tinha na aula | 1 |
215,829 | 7,298,320,298 | IssuesEvent | 2018-02-26 16:38:05 | escaladesports/esca-scripts | https://api.github.com/repos/escaladesports/esca-scripts | closed | Reset | high priority | Update keywords with split title.
Update any occurrence of the title in package.json with new title. | 1.0 | Reset - Update keywords with split title.
Update any occurrence of the title in package.json with new title. | non_test | reset update keywords with split title update any occurrence of the title in package json with new title | 0 |
11,273 | 3,195,471,157 | IssuesEvent | 2015-09-30 16:43:11 | ntop/ntopng | https://api.github.com/repos/ntop/ntopng | closed | Preferences: Interface Speed issue. | enhancement Testing Needed | If I use, as input interface, a ZC "virtual" interface or an aggregated "view" interface I can't choose an "Interface Speed" value bigger than 1000 Mbit.
In this example the view interface "view:zc:2@0,zc:2@1,zc:2@2,zc:2@3" represents a tap interface of a 1Gbit/s full duplex link. So the Interface Speed should be 2000Mbit.

I've the same problem with a ZC interface obtained as a "fan-out" (zbalance_ipc command with the option -m 2) of more than two 1Gbit/s interfaces. Also in this case I can't choose an "Interface Speed" value bigger than 1000 Mbit.
Thanks.
| 1.0 | Preferences: Interface Speed issue. - If I use, as input interface, a ZC "virtual" interface or an aggregated "view" interface I can't choose an "Interface Speed" value bigger than 1000 Mbit.
In this example the view interface "view:zc:2@0,zc:2@1,zc:2@2,zc:2@3" represents a tap interface of a 1Gbit/s full duplex link. So the Interface Speed should be 2000Mbit.

I've the same problem with a ZC interface obtained as a "fan-out" (zbalance_ipc command with the option -m 2) of more than two 1Gbit/s interfaces. Also in this case I can't choose an "Interface Speed" value bigger than 1000 Mbit.
Thanks.
| test | preferences interface speed issue if i use as input interface a zc virtual interface or an aggregated view interface i can t choose an interface speed value bigger than mbit in this example the view interface view zc zc zc zc represents a tap interface of a s full duplex link so the interface speed should be i ve the same problem with a zc interface obtained as a fan out zbalance ipc command with the option m of more than two s interfaces also in this case i can t choose an interface speed value bigger than mbit thanks | 1 |
324,122 | 23,984,553,306 | IssuesEvent | 2022-09-13 17:51:28 | pharmaverse/admiraldev | https://api.github.com/repos/pharmaverse/admiraldev | closed | Documentation: `assert_function_param` is in need of roxygen documentation | documentation | ### Please select a category the issue is focused on?
_No response_
### Let us know where something needs a refresh or put your idea here!

| 1.0 | Documentation: `assert_function_param` is in need of roxygen documentation - ### Please select a category the issue is focused on?
_No response_
### Let us know where something needs a refresh or put your idea here!

| non_test | documentation assert function param is in need of roxygen documentation please select a category the issue is focused on no response let us know where something needs a refresh or put your idea here | 0 |
43,819 | 23,380,334,960 | IssuesEvent | 2022-08-11 08:52:24 | tokio-rs/axum | https://api.github.com/repos/tokio-rs/axum | closed | Add benchmarks | E-help-wanted C-performance | The recent discussions around the [TechEmpower benchmarks](https://github.com/tokio-rs/axum/issues/1177) showed that, while axum is plenty fast, having some benchmarks in the repo would be useful:
- The purpose of the benchmarks should be to compare axum to itself such that one can easily see if some change improves performance.
- I don't they should do git shenanigans to automatically compare to a previous commit. Just print numbers for the current commit.
- They should not be designed to compare axum to other frameworks. There are other benchmarks that do that.
- They should be easy to run. Ideally just `cargo bench` if the output is good enough. Some bash script using `rewrk` is also fine with me.
- They should exercise things most users will use like routing and JSON input/output. Non axum specifics, like databases, aren't relevant imo.
- It's important that the benchmarks don't use unrealistic code. They should use the APIs we expect users to use.
- I don't necessarily think we should run the benchmarks on CI but should at least type check them to prevent bit rot.
I don't intend to chase micro benchmarks but having some convenient scripts to run does have value imo. | True | Add benchmarks - The recent discussions around the [TechEmpower benchmarks](https://github.com/tokio-rs/axum/issues/1177) showed that, while axum is plenty fast, having some benchmarks in the repo would be useful:
- The purpose of the benchmarks should be to compare axum to itself such that one can easily see if some change improves performance.
- I don't they should do git shenanigans to automatically compare to a previous commit. Just print numbers for the current commit.
- They should not be designed to compare axum to other frameworks. There are other benchmarks that do that.
- They should be easy to run. Ideally just `cargo bench` if the output is good enough. Some bash script using `rewrk` is also fine with me.
- They should exercise things most users will use like routing and JSON input/output. Non axum specifics, like databases, aren't relevant imo.
- It's important that the benchmarks don't use unrealistic code. They should use the APIs we expect users to use.
- I don't necessarily think we should run the benchmarks on CI but should at least type check them to prevent bit rot.
I don't intend to chase micro benchmarks but having some convenient scripts to run does have value imo. | non_test | add benchmarks the recent discussions around the showed that while axum is plenty fast having some benchmarks in the repo would be useful the purpose of the benchmarks should be to compare axum to itself such that one can easily see if some change improves performance i don t they should do git shenanigans to automatically compare to a previous commit just print numbers for the current commit they should not be designed to compare axum to other frameworks there are other benchmarks that do that they should be easy to run ideally just cargo bench if the output is good enough some bash script using rewrk is also fine with me they should exercise things most users will use like routing and json input output non axum specifics like databases aren t relevant imo it s important that the benchmarks don t use unrealistic code they should use the apis we expect users to use i don t necessarily think we should run the benchmarks on ci but should at least type check them to prevent bit rot i don t intend to chase micro benchmarks but having some convenient scripts to run does have value imo | 0 |
331,129 | 28,508,083,309 | IssuesEvent | 2023-04-19 00:07:19 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | `What's new` page appears as defaulted tab, before even `brave://welcome`, upon Griffin-seed restart | bug QA/Yes QA/Test-Plan-Specified OS/Desktop | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
`What's new` page appears as defaulted tab, before even `brave://welcome`, upon Griffin-seed restart
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. install `1.52.42`
2. launch Brave using `--variations-server-url=https://variations.bravesoftware.com/seed`
3. shut down
4. relaunch using `--variations-server-url=https://variations.bravesoftware.com/seed`
## Actual result:
<!--Please add screenshots if needed-->
example | example | example
---------|----------|--------
<img width="1312" alt="Screen Shot 2023-04-13 at 12 25 46 PM" src="https://user-images.githubusercontent.com/387249/231871853-237846ab-f1d2-4cc0-a7d8-df9d2f7cdff1.png"> | <img width="1312" alt="Screen Shot 2023-04-13 at 1 14 34 PM" src="https://user-images.githubusercontent.com/387249/231873122-32828b1f-719f-4533-b6cb-314c2d50f779.png"> | <img width="1379" alt="Screen Shot 2023-04-13 at 11 58 51 AM" src="https://user-images.githubusercontent.com/387249/231873126-f2c33fc6-048e-4e30-ba1a-df01827ac445.png">
## Expected result:
Not sure, but as the user hasn't even gone through the onboarding/`brave://welcome` experience yet, we shouldn't show this (yet/here)
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100%
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.52.42 Chromium: 112.0.5615.87 (Official Build) nightly (x86_64)
-- | --
Revision | 12315abe04d2dc2cac3055bd52ce662b2e742fbe-refs/branch-heads/5615_49@{#4}
OS | macOS Version 11.7.6 (Build 20G1231)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release?
- Can you reproduce this issue with the beta channel?
- Can you reproduce this issue with the nightly channel?
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
/cc @rebron @simonhong @bsclifton @kjozwiak @LaurenWags @aguscruiz | 1.0 | `What's new` page appears as defaulted tab, before even `brave://welcome`, upon Griffin-seed restart - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
`What's new` page appears as defaulted tab, before even `brave://welcome`, upon Griffin-seed restart
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. install `1.52.42`
2. launch Brave using `--variations-server-url=https://variations.bravesoftware.com/seed`
3. shut down
4. relaunch using `--variations-server-url=https://variations.bravesoftware.com/seed`
## Actual result:
<!--Please add screenshots if needed-->
example | example | example
---------|----------|--------
<img width="1312" alt="Screen Shot 2023-04-13 at 12 25 46 PM" src="https://user-images.githubusercontent.com/387249/231871853-237846ab-f1d2-4cc0-a7d8-df9d2f7cdff1.png"> | <img width="1312" alt="Screen Shot 2023-04-13 at 1 14 34 PM" src="https://user-images.githubusercontent.com/387249/231873122-32828b1f-719f-4533-b6cb-314c2d50f779.png"> | <img width="1379" alt="Screen Shot 2023-04-13 at 11 58 51 AM" src="https://user-images.githubusercontent.com/387249/231873126-f2c33fc6-048e-4e30-ba1a-df01827ac445.png">
## Expected result:
Not sure, but as the user hasn't even gone through the onboarding/`brave://welcome` experience yet, we shouldn't show this (yet/here)
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100%
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.52.42 Chromium: 112.0.5615.87 (Official Build) nightly (x86_64)
-- | --
Revision | 12315abe04d2dc2cac3055bd52ce662b2e742fbe-refs/branch-heads/5615_49@{#4}
OS | macOS Version 11.7.6 (Build 20G1231)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release?
- Can you reproduce this issue with the beta channel?
- Can you reproduce this issue with the nightly channel?
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
/cc @rebron @simonhong @bsclifton @kjozwiak @LaurenWags @aguscruiz | test | what s new page appears as defaulted tab before even brave welcome upon griffin seed restart have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description what s new page appears as defaulted tab before even brave welcome upon griffin seed restart steps to reproduce install launch brave using variations server url shut down relaunch using variations server url actual result example example example img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at am src expected result not sure but as the user hasn t even gone through the onboarding brave welcome experience yet we shouldn t show this yet here reproduces how often brave version brave version info brave chromium official build nightly revision refs branch heads os macos version build version channel information can you reproduce this issue with the current release can you reproduce this issue with the beta channel can you reproduce this issue with the nightly channel other additional information does the issue resolve itself when disabling brave shields does the issue resolve itself when disabling brave rewards is the issue reproducible on the latest version of chrome miscellaneous information cc rebron simonhong bsclifton kjozwiak laurenwags aguscruiz | 1 |
102,569 | 4,156,790,243 | IssuesEvent | 2016-06-16 19:07:02 | tsgrp/HPI | https://api.github.com/repos/tsgrp/HPI | closed | Cannot attach a controlled document past version 1.0 to a form and route for approval | Alfresco Controlled Doc Solution High Priority issue | Now that a chron id structure has been developed for our Alfresco implementation, findFirstVersionId returns the chron id instead of simply the passed in objectId (see OC revision 14034). Workflow documents are getting attached incorrectly past 1.0. | 1.0 | Cannot attach a controlled document past version 1.0 to a form and route for approval - Now that a chron id structure has been developed for our Alfresco implementation, findFirstVersionId returns the chron id instead of simply the passed in objectId (see OC revision 14034). Workflow documents are getting attached incorrectly past 1.0. | non_test | cannot attach a controlled document past version to a form and route for approval now that a chron id structure has been developed for our alfresco implementation findfirstversionid returns the chron id instead of simply the passed in objectid see oc revision workflow documents are getting attached incorrectly past | 0 |
24,303 | 12,258,313,012 | IssuesEvent | 2020-05-06 14:56:22 | flutter/flutter | https://api.github.com/repos/flutter/flutter | opened | regression in animated_placeholder_perf average_frame_build_time_millis | engine perf: speed severe: performance severe: regression team: benchmark ⚠ TODAY | Looks like https://github.com/flutter/flutter/pull/56378 is the culprit.

| True | regression in animated_placeholder_perf average_frame_build_time_millis - Looks like https://github.com/flutter/flutter/pull/56378 is the culprit.

| non_test | regression in animated placeholder perf average frame build time millis looks like is the culprit | 0 |
125,811 | 10,354,343,729 | IssuesEvent | 2019-09-05 13:34:47 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: tpchbench/tpch/nodes=3/cpu=4/sf=1 failed | C-test-failure O-roachtest O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/96b1500e20575ee5c609a00857c78c918078c99b
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=tpchbench/tpch/nodes=3/cpu=4/sf=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1465459&tab=buildLog
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190904-1465459/tpchbench/tpch/nodes=3/cpu=4/sf=1/run_1
tpchbench.go:123,cluster.go:2091,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1567578879-16-n4cpu4:4 -- ./workload run querybench --db=tpch --concurrency=1 --query-file=tpch --num-runs=3 --max-ops=63 --vectorize= {pgurl:1-3} --histograms=perf/stats.json --histograms-max-latency=8m20s returned:
stderr:
stdout:
0 AND l_shipmode IN ('AIR', 'AIR REG') AND l_shipinstruct = 'DELIVER IN PERSON') OR ( p_partkey = l_partkey AND p_brand = 'Brand#34' AND p_container IN ('LG CASE', 'LG BOX', 'LG PACK', 'LG PKG') AND l_quantity >= 20 AND l_quantity <= 20 + 10 AND p_size BETWEEN 1 AND 15 AND l_shipmode IN ('AIR', 'AIR REG') AND l_shipinstruct = 'DELIVER IN PERSON')
412.0s 0 0.0 0.0 0.0 0.0 0.0 0.0 19: SELECT s_name, s_address FROM supplier, nation WHERE s_suppkey IN ( SELECT ps_suppkey FROM partsupp WHERE ps_partkey IN ( SELECT p_partkey FROM part WHERE p_name LIKE 'forest%') AND ps_availqty > ( SELECT 0.5 * sum(l_quantity) FROM lineitem WHERE l_partkey = ps_partkey AND l_suppkey = ps_suppkey AND l_shipdate >= DATE '1994-01-01' AND l_shipdate < DATE '1994-01-01' + INTERVAL '1' YEAR)) AND s_nationkey = n_nationkey AND n_name = 'CANADA' ORDER BY s_name
Error: pq: internal error: unexpected error from the vectorized runtime: runtime error: index out of range
Error: exit status 1
: exit status 1
``` | 2.0 | roachtest: tpchbench/tpch/nodes=3/cpu=4/sf=1 failed - SHA: https://github.com/cockroachdb/cockroach/commits/96b1500e20575ee5c609a00857c78c918078c99b
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=tpchbench/tpch/nodes=3/cpu=4/sf=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1465459&tab=buildLog
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190904-1465459/tpchbench/tpch/nodes=3/cpu=4/sf=1/run_1
tpchbench.go:123,cluster.go:2091,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1567578879-16-n4cpu4:4 -- ./workload run querybench --db=tpch --concurrency=1 --query-file=tpch --num-runs=3 --max-ops=63 --vectorize= {pgurl:1-3} --histograms=perf/stats.json --histograms-max-latency=8m20s returned:
stderr:
stdout:
0 AND l_shipmode IN ('AIR', 'AIR REG') AND l_shipinstruct = 'DELIVER IN PERSON') OR ( p_partkey = l_partkey AND p_brand = 'Brand#34' AND p_container IN ('LG CASE', 'LG BOX', 'LG PACK', 'LG PKG') AND l_quantity >= 20 AND l_quantity <= 20 + 10 AND p_size BETWEEN 1 AND 15 AND l_shipmode IN ('AIR', 'AIR REG') AND l_shipinstruct = 'DELIVER IN PERSON')
412.0s 0 0.0 0.0 0.0 0.0 0.0 0.0 19: SELECT s_name, s_address FROM supplier, nation WHERE s_suppkey IN ( SELECT ps_suppkey FROM partsupp WHERE ps_partkey IN ( SELECT p_partkey FROM part WHERE p_name LIKE 'forest%') AND ps_availqty > ( SELECT 0.5 * sum(l_quantity) FROM lineitem WHERE l_partkey = ps_partkey AND l_suppkey = ps_suppkey AND l_shipdate >= DATE '1994-01-01' AND l_shipdate < DATE '1994-01-01' + INTERVAL '1' YEAR)) AND s_nationkey = n_nationkey AND n_name = 'CANADA' ORDER BY s_name
Error: pq: internal error: unexpected error from the vectorized runtime: runtime error: index out of range
Error: exit status 1
: exit status 1
``` | test | roachtest tpchbench tpch nodes cpu sf failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests tpchbench tpch nodes cpu sf pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts tpchbench tpch nodes cpu sf run tpchbench go cluster go errgroup go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run querybench db tpch concurrency query file tpch num runs max ops vectorize pgurl histograms perf stats json histograms max latency returned stderr stdout and l shipmode in air air reg and l shipinstruct deliver in person or p partkey l partkey and p brand brand and p container in lg case lg box lg pack lg pkg and l quantity and l quantity and p size between and and l shipmode in air air reg and l shipinstruct deliver in person select s name s address from supplier nation where s suppkey in select ps suppkey from partsupp where ps partkey in select p partkey from part where p name like forest and ps availqty select sum l quantity from lineitem where l partkey ps partkey and l suppkey ps suppkey and l shipdate date and l shipdate date interval year and s nationkey n nationkey and n name canada order by s name error pq internal error unexpected error from the vectorized runtime runtime error index out of range error exit status exit status | 1 |
334,375 | 29,832,967,475 | IssuesEvent | 2023-06-18 13:50:52 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix general.test_all_equal | Sub Task Failing Test | | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix general.test_all_equal - | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5285391556/jobs/9563837520"><img src=https://img.shields.io/badge/-success-success></a>
| test | fix general test all equal numpy a href src torch a href src tensorflow a href src jax a href src paddle a href src | 1 |
168,368 | 13,081,992,565 | IssuesEvent | 2020-08-01 13:00:54 | ValveSoftware/steam-for-linux | https://api.github.com/repos/ValveSoftware/steam-for-linux | closed | How and where install missings libraries ? | 3rd party game Distro Family: Arch Need Retest | #### Your system information
* Steam client version (build number or date):
2 april 2018
* Distribution (e.g. Ubuntu):
archlinux
* Opted into Steam client beta?: [Yes/No]
no (i think...)
* Have you checked for system updates?: [Yes/No]
yes
#### Please describe your issue in as much detail as possible:
So, I have some games who can't start (crash on start) After reading some topics, it seem that is because of some missing libraries.
I launched this command line into ´$HOME/.local/share/Steam/ubuntu12_64´
```
ldd $(file *|sed '/ELF/!d;s/:.*//g')|grep 'not found'|sort|uniq
```
```
libavcodec.so.57 => not found
libavformat.so.57 => not found
libavresample.so.3 => not found
libavutil.so.55 => not found
libswscale.so.4 => not found
libtier0_s.so => not found
libv8.so => not found
libvideo.so => not found
libvstdlib_s.so => not found
```
so, ffmpeg is already installed, and should contain some of this libraries... so, I do not understand something... Somebody can help me ? | 1.0 | How and where install missings libraries ? - #### Your system information
* Steam client version (build number or date):
2 april 2018
* Distribution (e.g. Ubuntu):
archlinux
* Opted into Steam client beta?: [Yes/No]
no (i think...)
* Have you checked for system updates?: [Yes/No]
yes
#### Please describe your issue in as much detail as possible:
So, I have some games who can't start (crash on start) After reading some topics, it seem that is because of some missing libraries.
I launched this command line into ´$HOME/.local/share/Steam/ubuntu12_64´
```
ldd $(file *|sed '/ELF/!d;s/:.*//g')|grep 'not found'|sort|uniq
```
```
libavcodec.so.57 => not found
libavformat.so.57 => not found
libavresample.so.3 => not found
libavutil.so.55 => not found
libswscale.so.4 => not found
libtier0_s.so => not found
libv8.so => not found
libvideo.so => not found
libvstdlib_s.so => not found
```
so, ffmpeg is already installed, and should contain some of this libraries... so, I do not understand something... Somebody can help me ? | test | how and where install missings libraries your system information steam client version build number or date april distribution e g ubuntu archlinux opted into steam client beta no i think have you checked for system updates yes please describe your issue in as much detail as possible so i have some games who can t start crash on start after reading some topics it seem that is because of some missing libraries i launched this command line into ´ home local share steam ´ ldd file sed elf d s g grep not found sort uniq libavcodec so not found libavformat so not found libavresample so not found libavutil so not found libswscale so not found s so not found so not found libvideo so not found libvstdlib s so not found so ffmpeg is already installed and should contain some of this libraries so i do not understand something somebody can help me | 1 |
153,969 | 12,179,119,497 | IssuesEvent | 2020-04-28 10:07:41 | spring-projects/spring-framework | https://api.github.com/repos/spring-projects/spring-framework | closed | Upgrade to JUnit 4.13 | in: test type: dependency-upgrade | ## Overview
JUnit 4.13 will be released soon. We should therefore verify that the Spring Framework build and especially the `spring-test` module work properly with JUnit 4.13 before it goes GA and then upgrade to 4.13 once it is officially released.
## Deliverables
- [ ] Upgrade to JUnit 4.13 as soon as it is [released](https://github.com/junit-team/junit4/issues/1496).
- [x] Upgrade to JUnit 4.13-beta-3 in the interim.
| 1.0 | Upgrade to JUnit 4.13 - ## Overview
JUnit 4.13 will be released soon. We should therefore verify that the Spring Framework build and especially the `spring-test` module work properly with JUnit 4.13 before it goes GA and then upgrade to 4.13 once it is officially released.
## Deliverables
- [ ] Upgrade to JUnit 4.13 as soon as it is [released](https://github.com/junit-team/junit4/issues/1496).
- [x] Upgrade to JUnit 4.13-beta-3 in the interim.
| test | upgrade to junit overview junit will be released soon we should therefore verify that the spring framework build and especially the spring test module work properly with junit before it goes ga and then upgrade to once it is officially released deliverables upgrade to junit as soon as it is upgrade to junit beta in the interim | 1 |
175,481 | 21,313,836,875 | IssuesEvent | 2022-04-16 01:06:33 | Nivaskumark/kernel_v4.1.15 | https://api.github.com/repos/Nivaskumark/kernel_v4.1.15 | opened | CVE-2017-17863 (High) detected in linuxlinux-4.6 | security vulnerability | ## CVE-2017-17863 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
kernel/bpf/verifier.c in the Linux kernel 4.9.x through 4.9.71 does not check the relationship between pointer values and the BPF stack, which allows local users to cause a denial of service (integer overflow or invalid memory access) or possibly have unspecified other impact.
<p>Publish Date: 2017-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-17863>CVE-2017-17863</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-17863">https://nvd.nist.gov/vuln/detail/CVE-2017-17863</a></p>
<p>Release Date: 2017-12-27</p>
<p>Fix Resolution: linux-hardened - 4.14.8.b-1;linux - 4.14.8-1;linux-lts - 4.9.69-1;linux-zen - 4.14.8-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-17863 (High) detected in linuxlinux-4.6 - ## CVE-2017-17863 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
kernel/bpf/verifier.c in the Linux kernel 4.9.x through 4.9.71 does not check the relationship between pointer values and the BPF stack, which allows local users to cause a denial of service (integer overflow or invalid memory access) or possibly have unspecified other impact.
<p>Publish Date: 2017-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-17863>CVE-2017-17863</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-17863">https://nvd.nist.gov/vuln/detail/CVE-2017-17863</a></p>
<p>Release Date: 2017-12-27</p>
<p>Fix Resolution: linux-hardened - 4.14.8.b-1;linux - 4.14.8-1;linux-lts - 4.9.69-1;linux-zen - 4.14.8-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files kernel bpf verifier c kernel bpf verifier c vulnerability details kernel bpf verifier c in the linux kernel x through does not check the relationship between pointer values and the bpf stack which allows local users to cause a denial of service integer overflow or invalid memory access or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux hardened b linux linux lts linux zen step up your open source security game with whitesource | 0 |
198,375 | 6,972,724,798 | IssuesEvent | 2017-12-11 17:58:13 | osuosl/streamwebs | https://api.github.com/repos/osuosl/streamwebs | closed | Create manage account page (mockup in account_mockup branch) that is accessible by admin for specific org and is generally accessible by the super admin. | high priority Needs Implementation | Acceptance Criteria
Implement what you see on story board with the following changes:
- [x] Implement all buttons as different named submit buttons to same form action (a.k.a. View Method)
- [x] Change Author Accounts to Contributors
- [x] Change Admin Accounts to Editors
- [x] Change Approve New Users columns: Admin -> Editor, Student -> Contributor
- [x] Remove Birthday Column from grids
- [x] When you click Apply, apply appropriate permissions and redisplay the view (apply is a submit form button)
- [x] Contributor, Editor and Deny check boxes should be mutually exclusive
- [x] When you click on a All check box JS selects/deselects all check boxes in the grid
- [x] When you click on Promote, any selected users are promoted to Editor accounts.
- [x] When you click on Demote, does opposite as Promote
- [x] When you click on Delete Selected, selected accounts in the corresponding grid are removed permanently (beware FK issues here)
- [x] Sections for the various grids should be clearly delineated
- [x] Change Promote/Demote to include the word Selected
- [x] Make Contributors last on the page since it will be the biggest list
| 1.0 | Create manage account page (mockup in account_mockup branch) that is accessible by admin for specific org and is generally accessible by the super admin. - Acceptance Criteria
Implement what you see on story board with the following changes:
- [x] Implement all buttons as different named submit buttons to same form action (a.k.a. View Method)
- [x] Change Author Accounts to Contributors
- [x] Change Admin Accounts to Editors
- [x] Change Approve New Users columns: Admin -> Editor, Student -> Contributor
- [x] Remove Birthday Column from grids
- [x] When you click Apply, apply appropriate permissions and redisplay the view (apply is a submit form button)
- [x] Contributor, Editor and Deny check boxes should be mutually exclusive
- [x] When you click on a All check box JS selects/deselects all check boxes in the grid
- [x] When you click on Promote, any selected users are promoted to Editor accounts.
- [x] When you click on Demote, does opposite as Promote
- [x] When you click on Delete Selected, selected accounts in the corresponding grid are removed permanently (beware FK issues here)
- [x] Sections for the various grids should be clearly delineated
- [x] Change Promote/Demote to include the word Selected
- [x] Make Contributors last on the page since it will be the biggest list
| non_test | create manage account page mockup in account mockup branch that is accessible by admin for specific org and is generally accessible by the super admin acceptance criteria implement what you see on story board with the following changes implement all buttons as different named submit buttons to same form action a k a view method change author accounts to contributors change admin accounts to editors change approve new users columns admin editor student contributor remove birthday column from grids when you click apply apply appropriate permissions and redisplay the view apply is a submit form button contributor editor and deny check boxes should be mutually exclusive when you click on a all check box js selects deselects all check boxes in the grid when you click on promote any selected users are promoted to editor accounts when you click on demote does opposite as promote when you click on delete selected selected accounts in the corresponding grid are removed permanently beware fk issues here sections for the various grids should be clearly delineated change promote demote to include the word selected make contributors last on the page since it will be the biggest list | 0 |
35,966 | 4,707,034,677 | IssuesEvent | 2016-10-13 18:55:02 | Automattic/wp-calypso | https://api.github.com/repos/Automattic/wp-calypso | closed | Billing History: "you have no saved cards" message isn't styled consistently | Billing Me [Status] Needs Design Review [Type] Bug | #### Steps to reproduce
1. Starting at URL: https://wpcalypso.wordpress.com/me/billing
2. When no credit cards are saved
3. Look at "Manage Your Credit Cards" section
#### What I expected
To see the message styled as other "empty" message in the same page.
<img width="755" alt="image23" src="https://cloud.githubusercontent.com/assets/66797/19173364/d370ea5e-8bdb-11e6-9e39-c00541c71c0e.png">
#### Context
Shared by Beau in a recent #dogfooding session (I confirmed in Chrome on OS X). | 1.0 | Billing History: "you have no saved cards" message isn't styled consistently - #### Steps to reproduce
1. Starting at URL: https://wpcalypso.wordpress.com/me/billing
2. When no credit cards are saved
3. Look at "Manage Your Credit Cards" section
#### What I expected
To see the message styled as other "empty" message in the same page.
<img width="755" alt="image23" src="https://cloud.githubusercontent.com/assets/66797/19173364/d370ea5e-8bdb-11e6-9e39-c00541c71c0e.png">
#### Context
Shared by Beau in a recent #dogfooding session (I confirmed in Chrome on OS X). | non_test | billing history you have no saved cards message isn t styled consistently steps to reproduce starting at url when no credit cards are saved look at manage your credit cards section what i expected to see the message styled as other empty message in the same page img width alt src context shared by beau in a recent dogfooding session i confirmed in chrome on os x | 0 |
81,793 | 31,682,205,523 | IssuesEvent | 2023-09-08 01:30:29 | dotCMS/core | https://api.github.com/repos/dotCMS/core | closed | CLI Workflow is missing the systemActionMappings attribute | Type : Defect Merged QA : Needs Internal Team : Scout Triage dotCMS : CLI | ### Parent Issue
_No response_
### Problem Statement
We should support systemActionMappings as they are essential part of Workflows and our models don't know anything about that attribute
### Steps to Reproduce
Pull a CT using the cli e.g. FileAsset
Now compare it against this https://demo.dotcms.com/api/v1/contenttype/id/fileAsset
notice the file pulled using the CLI does not have that attribute
### Acceptance Criteria
The files should both martch
### dotCMS Version
current master
### Proposed Objective
Core Features
### Proposed Priority
Priority 2 - Important
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | 1.0 | CLI Workflow is missing the systemActionMappings attribute - ### Parent Issue
_No response_
### Problem Statement
We should support systemActionMappings as they are essential part of Workflows and our models don't know anything about that attribute
### Steps to Reproduce
Pull a CT using the cli e.g. FileAsset
Now compare it against this https://demo.dotcms.com/api/v1/contenttype/id/fileAsset
notice the file pulled using the CLI does not have that attribute
### Acceptance Criteria
The files should both martch
### dotCMS Version
current master
### Proposed Objective
Core Features
### Proposed Priority
Priority 2 - Important
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | non_test | cli workflow is missing the systemactionmappings attribute parent issue no response problem statement we should support systemactionmappings as they are essential part of workflows and our models don t know anything about that attribute steps to reproduce pull a ct using the cli e g fileasset now compare it against this notice the file pulled using the cli does not have that attribute acceptance criteria the files should both martch dotcms version current master proposed objective core features proposed priority priority important external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response quality assurance notes workarounds no response sub tasks estimates no response | 0 |
184,125 | 14,969,887,113 | IssuesEvent | 2021-01-27 18:47:57 | bridgetownrb/bridgetown | https://api.github.com/repos/bridgetownrb/bridgetown | closed | docs: Add information about how to customize PostCSS | documentation | With the PostCSS config that's landed in main, I'd like to reference useful documentation on how to configure it to support various features…nesting, custom media query variables, environment variables, etc. It looks like preset-env provides a lot of functionality out of the box, but it probably requires changing the "stage: 3" config to something else or adding an allowlist of desired features. There may be additional plugins out there that are popular as well we could link to.
ref: http://preset-env.cssdb.org/
https://github.com/csstools/postcss-preset-env#features
https://github.com/csstools/postcss-preset-env/blob/master/src/lib/plugins-by-id.js#L36 | 1.0 | docs: Add information about how to customize PostCSS - With the PostCSS config that's landed in main, I'd like to reference useful documentation on how to configure it to support various features…nesting, custom media query variables, environment variables, etc. It looks like preset-env provides a lot of functionality out of the box, but it probably requires changing the "stage: 3" config to something else or adding an allowlist of desired features. There may be additional plugins out there that are popular as well we could link to.
ref: http://preset-env.cssdb.org/
https://github.com/csstools/postcss-preset-env#features
https://github.com/csstools/postcss-preset-env/blob/master/src/lib/plugins-by-id.js#L36 | non_test | docs add information about how to customize postcss with the postcss config that s landed in main i d like to reference useful documentation on how to configure it to support various features…nesting custom media query variables environment variables etc it looks like preset env provides a lot of functionality out of the box but it probably requires changing the stage config to something else or adding an allowlist of desired features there may be additional plugins out there that are popular as well we could link to ref | 0 |
146,710 | 5,626,320,199 | IssuesEvent | 2017-04-04 21:33:26 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | TestShortNames {storage} | kind/flake priority/P2 | https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/4318/
Failed: TestShortNames {storage}
```
/usr/local/go/src/runtime/panic.go:566 +0x95 fp=0xc4206350b8 sp=0xc420635098
/usr/local/go/src/runtime/sigpanic_unix.go:27 +0x288 fp=0xc420635110 sp=0xc4206350b8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/node.go:205 +0x195 fp=0xc4206352a8 sp=0xc420635110
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/bucket.go:598 +0x11f fp=0xc420635308 sp=0xc4206352a8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/bucket.go:181 +0x545 fp=0xc420635460 sp=0xc420635308
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/tx.go:108 +0x61 fp=0xc4206354c0 sp=0xc420635460
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/backend/batch_tx.go:53 +0xa4 fp=0xc420635570 sp=0xc4206354c0
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/kvstore.go:116 +0x414 fp=0xc420635638 sp=0xc420635570
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/watchable_store.go:73 +0x92 fp=0xc4206356d8 sp=0xc420635638
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/watchable_store.go:68 +0x6f fp=0xc420635728 sp=0xc4206356d8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/etcdserver/server.go:405 +0x17c0 fp=0xc420635cc0 sp=0xc420635728
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:558 +0x1d1 fp=0xc420635f58 sp=0xc420635cc0
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:159 +0x39 fp=0xc420635f90 sp=0xc420635f58
/usr/local/go/src/runtime/asm_amd64.s:2086 +0x1 fp=0xc420635f98 sp=0xc420635f90
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:160 +0xe9
/usr/local/go/src/testing/testing.go:647 +0x56e
/usr/local/go/src/testing/testing.go:793 +0xba
/usr/local/go/src/testing/testing.go:610 +0xca
/usr/local/go/src/testing/testing.go:799 +0x4bb
/usr/local/go/src/testing/testing.go:743 +0x130
k8s.io/kubernetes/pkg/registry/core/configmap/storage/_test/_testmain.go:66 +0x1b9
/usr/local/go/src/runtime/asm_amd64.s:2086 +0x1
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/golang/glog/glog.go:879 +0x9e
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/golang/glog/glog.go:410 +0x23b
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:163 +0x17b
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:755 +0x13e
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/storage/etcd/testing/utils.go:316 +0xd0
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/pkg/registry/registrytest/etcd.go:41 +0x6d
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/pkg/registry/core/configmap/storage/storage_test.go:33 +0x63
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/pkg/registry/core/configmap/storage/storage_test.go:164 +0x3d
/usr/local/go/src/testing/testing.go:610 +0xca
/usr/local/go/src/testing/testing.go:646 +0x530
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:174 +0xb5
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:92 +0x125
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:174 +0xb5
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:92 +0x125
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/wal/file_pipeline.go:89 +0x24a
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/wal/file_pipeline.go:47 +0x1e8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/schedule/schedule.go:146 +0x4e4
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/schedule/schedule.go:71 +0x2e8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/lease/lessor.go:386 +0x1e8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/lease/lessor.go:171 +0x267
/usr/local/go/src/runtime/netpoll.go:160 +0x5e
/usr/local/go/src/net/fd_poll_runtime.go:73 +0x5b
/usr/local/go/src/net/fd_poll_runtime.go:78 +0x42
/usr/local/go/src/net/fd_unix.go:419 +0x2b8
/usr/local/go/src/net/unixsock_posix.go:158 +0x51
/usr/local/go/src/net/unixsock.go:229 +0x50
<autogenerated>:91 +0x69
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/bridge.go:89 +0x91
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/bridge.go:53 +0x466
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/backend/backend.go:193 +0x26b
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/backend/backend.go:119 +0x29e
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/raft/node.go:307 +0x12cf
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/raft/node.go:204 +0x800
```
| 1.0 | TestShortNames {storage} - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/4318/
Failed: TestShortNames {storage}
```
/usr/local/go/src/runtime/panic.go:566 +0x95 fp=0xc4206350b8 sp=0xc420635098
/usr/local/go/src/runtime/sigpanic_unix.go:27 +0x288 fp=0xc420635110 sp=0xc4206350b8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/node.go:205 +0x195 fp=0xc4206352a8 sp=0xc420635110
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/bucket.go:598 +0x11f fp=0xc420635308 sp=0xc4206352a8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/bucket.go:181 +0x545 fp=0xc420635460 sp=0xc420635308
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/boltdb/bolt/tx.go:108 +0x61 fp=0xc4206354c0 sp=0xc420635460
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/backend/batch_tx.go:53 +0xa4 fp=0xc420635570 sp=0xc4206354c0
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/kvstore.go:116 +0x414 fp=0xc420635638 sp=0xc420635570
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/watchable_store.go:73 +0x92 fp=0xc4206356d8 sp=0xc420635638
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/watchable_store.go:68 +0x6f fp=0xc420635728 sp=0xc4206356d8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/etcdserver/server.go:405 +0x17c0 fp=0xc420635cc0 sp=0xc420635728
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:558 +0x1d1 fp=0xc420635f58 sp=0xc420635cc0
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:159 +0x39 fp=0xc420635f90 sp=0xc420635f58
/usr/local/go/src/runtime/asm_amd64.s:2086 +0x1 fp=0xc420635f98 sp=0xc420635f90
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:160 +0xe9
/usr/local/go/src/testing/testing.go:647 +0x56e
/usr/local/go/src/testing/testing.go:793 +0xba
/usr/local/go/src/testing/testing.go:610 +0xca
/usr/local/go/src/testing/testing.go:799 +0x4bb
/usr/local/go/src/testing/testing.go:743 +0x130
k8s.io/kubernetes/pkg/registry/core/configmap/storage/_test/_testmain.go:66 +0x1b9
/usr/local/go/src/runtime/asm_amd64.s:2086 +0x1
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/golang/glog/glog.go:879 +0x9e
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/golang/glog/glog.go:410 +0x23b
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:163 +0x17b
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/cluster.go:755 +0x13e
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/storage/etcd/testing/utils.go:316 +0xd0
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/pkg/registry/registrytest/etcd.go:41 +0x6d
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/pkg/registry/core/configmap/storage/storage_test.go:33 +0x63
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/pkg/registry/core/configmap/storage/storage_test.go:164 +0x3d
/usr/local/go/src/testing/testing.go:610 +0xca
/usr/local/go/src/testing/testing.go:646 +0x530
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:174 +0xb5
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:92 +0x125
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:174 +0xb5
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/logutil/merge_logger.go:92 +0x125
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/wal/file_pipeline.go:89 +0x24a
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/wal/file_pipeline.go:47 +0x1e8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/schedule/schedule.go:146 +0x4e4
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/pkg/schedule/schedule.go:71 +0x2e8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/lease/lessor.go:386 +0x1e8
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/lease/lessor.go:171 +0x267
/usr/local/go/src/runtime/netpoll.go:160 +0x5e
/usr/local/go/src/net/fd_poll_runtime.go:73 +0x5b
/usr/local/go/src/net/fd_poll_runtime.go:78 +0x42
/usr/local/go/src/net/fd_unix.go:419 +0x2b8
/usr/local/go/src/net/unixsock_posix.go:158 +0x51
/usr/local/go/src/net/unixsock.go:229 +0x50
<autogenerated>:91 +0x69
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/bridge.go:89 +0x91
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/integration/bridge.go:53 +0x466
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/backend/backend.go:193 +0x26b
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/mvcc/backend/backend.go:119 +0x29e
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/raft/node.go:307 +0x12cf
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/github.com/coreos/etcd/raft/node.go:204 +0x800
```
| non_test | testshortnames storage failed testshortnames storage usr local go src runtime panic go fp sp usr local go src runtime sigpanic unix go fp sp go src io kubernetes output local go src io kubernetes vendor github com boltdb bolt node go fp sp go src io kubernetes output local go src io kubernetes vendor github com boltdb bolt bucket go fp sp go src io kubernetes output local go src io kubernetes vendor github com boltdb bolt bucket go fp sp go src io kubernetes output local go src io kubernetes vendor github com boltdb bolt tx go fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd mvcc backend batch tx go fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd mvcc kvstore go fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd mvcc watchable store go fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd mvcc watchable store go fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd etcdserver server go fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd integration cluster go fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd integration cluster go fp sp usr local go src runtime asm s fp sp go src io kubernetes output local go src io kubernetes vendor github com coreos etcd integration cluster go usr local go src testing testing go usr local go src testing testing go usr local go src testing testing go usr local go src testing testing go usr local go src testing testing go io kubernetes pkg registry core configmap storage test testmain go usr local go src runtime asm s go src io kubernetes output local go src io kubernetes vendor github com golang glog glog go go src io kubernetes output local go src io kubernetes vendor github com golang glog glog go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd integration cluster go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd integration cluster go go src io kubernetes output local go src io kubernetes vendor io apiserver pkg storage etcd testing utils go go src io kubernetes output local go src io kubernetes pkg registry registrytest etcd go go src io kubernetes output local go src io kubernetes pkg registry core configmap storage storage test go go src io kubernetes output local go src io kubernetes pkg registry core configmap storage storage test go usr local go src testing testing go usr local go src testing testing go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd pkg logutil merge logger go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd pkg logutil merge logger go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd pkg logutil merge logger go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd pkg logutil merge logger go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd wal file pipeline go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd wal file pipeline go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd pkg schedule schedule go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd pkg schedule schedule go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd lease lessor go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd lease lessor go usr local go src runtime netpoll go usr local go src net fd poll runtime go usr local go src net fd poll runtime go usr local go src net fd unix go usr local go src net unixsock posix go usr local go src net unixsock go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd integration bridge go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd integration bridge go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd mvcc backend backend go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd mvcc backend backend go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd raft node go go src io kubernetes output local go src io kubernetes vendor github com coreos etcd raft node go | 0 |
40,726 | 21,075,111,577 | IssuesEvent | 2022-04-02 03:03:52 | haskell-unordered-containers/unordered-containers | https://api.github.com/repos/haskell-unordered-containers/unordered-containers | opened | `unionArrayBy` could be improved for some corner cases | performance | 1. https://github.com/haskell-unordered-containers/unordered-containers/blob/b6bde46eae679111c3f088964f192290cc08b6c8/Data/HashMap/Internal.hs#L1647-L1649
This could be used to improve the cases where we `union` a `BitmapIndexed` and a `Full` node.
2. For the case where we have two `Full` nodes, we don't need to use `countTrailingZeros` to find the next 1-bit.
Ultimately it would be nice (but not super important) if the perf regressions in the `union.Int` benchmark due to https://github.com/haskell-unordered-containers/unordered-containers/pull/395 could be undone.
| True | `unionArrayBy` could be improved for some corner cases - 1. https://github.com/haskell-unordered-containers/unordered-containers/blob/b6bde46eae679111c3f088964f192290cc08b6c8/Data/HashMap/Internal.hs#L1647-L1649
This could be used to improve the cases where we `union` a `BitmapIndexed` and a `Full` node.
2. For the case where we have two `Full` nodes, we don't need to use `countTrailingZeros` to find the next 1-bit.
Ultimately it would be nice (but not super important) if the perf regressions in the `union.Int` benchmark due to https://github.com/haskell-unordered-containers/unordered-containers/pull/395 could be undone.
| non_test | unionarrayby could be improved for some corner cases this could be used to improve the cases where we union a bitmapindexed and a full node for the case where we have two full nodes we don t need to use counttrailingzeros to find the next bit ultimately it would be nice but not super important if the perf regressions in the union int benchmark due to could be undone | 0 |
51,571 | 12,747,939,237 | IssuesEvent | 2020-06-26 19:00:13 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | opened | third_party/flatbuffers references missing filemirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip | type:build/install | **Describe the problem**
Some of our builds are failing:
> WARNING: Download from https://storage.googleapis.com/mirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip failed: class com.google.devtools.build.lib.bazel.repository.downloader.UnrecoverableHttpException GET returned 404 Not Found
The URL is from TensorFlow r2.2 https://github.com/tensorflow/tensorflow/blob/r2.2/third_party/flatbuffers/workspace.bzl#L11
**Provide the exact sequence of commands / steps that you executed before running into the problem**
```
$ wget "https://storage.googleapis.com/mirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip"
--2020-06-26 20:59:28-- https://storage.googleapis.com/mirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip
Resolving storage.googleapis.com (storage.googleapis.com)... 2a00:1450:4007:815::2010, 216.58.213.176, 172.217.22.144, ...
Connecting to storage.googleapis.com (storage.googleapis.com)|2a00:1450:4007:815::2010|:443... connected.
HTTP request sent, awaiting response... 404 Not Found
2020-06-26 20:59:29 ERROR 404: Not Found.
```
However, the second URL is fine:
```
wget "https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip"
--2020-06-26 20:59:50-- https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip
Resolving github.com (github.com)... 140.82.118.3
Connecting to github.com (github.com)|140.82.118.3|:443... connected.
HTTP request sent, awaiting response... 302 Found
Location: https://codeload.github.com/google/flatbuffers/zip/a4b2884e4ed6116335d534af8f58a84678b74a17 [following]
--2020-06-26 20:59:50-- https://codeload.github.com/google/flatbuffers/zip/a4b2884e4ed6116335d534af8f58a84678b74a17
Resolving codeload.github.com (codeload.github.com)... 140.82.112.10
Connecting to codeload.github.com (codeload.github.com)|140.82.112.10|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: unspecified [application/zip]
Saving to: 'a4b2884e4ed6116335d534af8f58a84678b74a17.zip'
a4b2884e4ed6116335d534af8f58a84678b74a17.zip [ <=> ] 564.91K 749KB/s ^
``` | 1.0 | third_party/flatbuffers references missing filemirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip - **Describe the problem**
Some of our builds are failing:
> WARNING: Download from https://storage.googleapis.com/mirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip failed: class com.google.devtools.build.lib.bazel.repository.downloader.UnrecoverableHttpException GET returned 404 Not Found
The URL is from TensorFlow r2.2 https://github.com/tensorflow/tensorflow/blob/r2.2/third_party/flatbuffers/workspace.bzl#L11
**Provide the exact sequence of commands / steps that you executed before running into the problem**
```
$ wget "https://storage.googleapis.com/mirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip"
--2020-06-26 20:59:28-- https://storage.googleapis.com/mirror.tensorflow.org/https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip
Resolving storage.googleapis.com (storage.googleapis.com)... 2a00:1450:4007:815::2010, 216.58.213.176, 172.217.22.144, ...
Connecting to storage.googleapis.com (storage.googleapis.com)|2a00:1450:4007:815::2010|:443... connected.
HTTP request sent, awaiting response... 404 Not Found
2020-06-26 20:59:29 ERROR 404: Not Found.
```
However, the second URL is fine:
```
wget "https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip"
--2020-06-26 20:59:50-- https://github.com/google/flatbuffers/archive/a4b2884e4ed6116335d534af8f58a84678b74a17.zip
Resolving github.com (github.com)... 140.82.118.3
Connecting to github.com (github.com)|140.82.118.3|:443... connected.
HTTP request sent, awaiting response... 302 Found
Location: https://codeload.github.com/google/flatbuffers/zip/a4b2884e4ed6116335d534af8f58a84678b74a17 [following]
--2020-06-26 20:59:50-- https://codeload.github.com/google/flatbuffers/zip/a4b2884e4ed6116335d534af8f58a84678b74a17
Resolving codeload.github.com (codeload.github.com)... 140.82.112.10
Connecting to codeload.github.com (codeload.github.com)|140.82.112.10|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: unspecified [application/zip]
Saving to: 'a4b2884e4ed6116335d534af8f58a84678b74a17.zip'
a4b2884e4ed6116335d534af8f58a84678b74a17.zip [ <=> ] 564.91K 749KB/s ^
``` | non_test | third party flatbuffers references missing filemirror tensorflow org describe the problem some of our builds are failing warning download from failed class com google devtools build lib bazel repository downloader unrecoverablehttpexception get returned not found the url is from tensorflow provide the exact sequence of commands steps that you executed before running into the problem wget resolving storage googleapis com storage googleapis com connecting to storage googleapis com storage googleapis com connected http request sent awaiting response not found error not found however the second url is fine wget resolving github com github com connecting to github com github com connected http request sent awaiting response found location resolving codeload github com codeload github com connecting to codeload github com codeload github com connected http request sent awaiting response ok length unspecified saving to zip zip s | 0 |
212,505 | 16,455,451,402 | IssuesEvent | 2021-05-21 11:56:06 | Reiningecho90/S.A.N.E.-AI | https://api.github.com/repos/Reiningecho90/S.A.N.E.-AI | closed | Error Out on File Load | bug in-work testing | New settings private test file is immediately erroring out upon attempting to load the file.
File should run to create tkinter GUI although that is not happening.
Windows 10 in VS Code | 1.0 | Error Out on File Load - New settings private test file is immediately erroring out upon attempting to load the file.
File should run to create tkinter GUI although that is not happening.
Windows 10 in VS Code | test | error out on file load new settings private test file is immediately erroring out upon attempting to load the file file should run to create tkinter gui although that is not happening windows in vs code | 1 |
315,295 | 27,063,080,383 | IssuesEvent | 2023-02-13 21:27:20 | ethyca/fides | https://api.github.com/repos/ethyca/fides | closed | Add a new Cypress E2E test runner that executes against a fully configured environment | enhancement Test Automation Improvements | ### Description
In some previous releases, we worked on building out test setup scripts (#1291) and a local manual test environment (#1292), which gives us the ability to do true E2E manual tests to confirm all features are working as expected when running together.
Our existing Cypress suites for the `admin-ui` and `privacy-center` are great, but they are isolated to their specific applications and tend to stub out their backends for performance. Therefore, there's still an open need for an E2E test runner to use for regression tests on an ongoing basis.
Once the runner is configured, let's put it to work by defining a core smoke test for releases. We should ensure this can run against our local test env first, but I want to immediately start using this against some hosted staging & demo environments in the short-term, so let's ensure that any URLs are easy to configure and override (e.g. http://localhost:3000 will quickly become something like https://fides.fides-staging.ethyca.com/)
In addition, let's signup for Cypress Cloud and publish results there, it's a great product!
### Acceptance Criteria
* MUST implement a new Cypress runner to run E2E tests, targeting the local "manual test" environment (see #1292)
* MUST support targeting different hosts via environment variables, so we can run against staging / demo / etc.
* MUST have a `nox` session (e.g. `nox -s e2e_test`) that runs the Cypress suite in headless mode
* MUST configure the test reporter to publish results to our Cypress Cloud account
* MUST write a single Cypress E2E test that runs a simple smoke test covering:
1. Confirm can login to the Admin UI
2. Confirm that the Postgres & Mongo connectors are configured
3. Confirm that the Privacy Center can be accessed
4. Submit an access request via the Privacy Center
5. Approve the access request via the Admin UI
6. Confirm the access request succeeds | 1.0 | Add a new Cypress E2E test runner that executes against a fully configured environment - ### Description
In some previous releases, we worked on building out test setup scripts (#1291) and a local manual test environment (#1292), which gives us the ability to do true E2E manual tests to confirm all features are working as expected when running together.
Our existing Cypress suites for the `admin-ui` and `privacy-center` are great, but they are isolated to their specific applications and tend to stub out their backends for performance. Therefore, there's still an open need for an E2E test runner to use for regression tests on an ongoing basis.
Once the runner is configured, let's put it to work by defining a core smoke test for releases. We should ensure this can run against our local test env first, but I want to immediately start using this against some hosted staging & demo environments in the short-term, so let's ensure that any URLs are easy to configure and override (e.g. http://localhost:3000 will quickly become something like https://fides.fides-staging.ethyca.com/)
In addition, let's signup for Cypress Cloud and publish results there, it's a great product!
### Acceptance Criteria
* MUST implement a new Cypress runner to run E2E tests, targeting the local "manual test" environment (see #1292)
* MUST support targeting different hosts via environment variables, so we can run against staging / demo / etc.
* MUST have a `nox` session (e.g. `nox -s e2e_test`) that runs the Cypress suite in headless mode
* MUST configure the test reporter to publish results to our Cypress Cloud account
* MUST write a single Cypress E2E test that runs a simple smoke test covering:
1. Confirm can login to the Admin UI
2. Confirm that the Postgres & Mongo connectors are configured
3. Confirm that the Privacy Center can be accessed
4. Submit an access request via the Privacy Center
5. Approve the access request via the Admin UI
6. Confirm the access request succeeds | test | add a new cypress test runner that executes against a fully configured environment description in some previous releases we worked on building out test setup scripts and a local manual test environment which gives us the ability to do true manual tests to confirm all features are working as expected when running together our existing cypress suites for the admin ui and privacy center are great but they are isolated to their specific applications and tend to stub out their backends for performance therefore there s still an open need for an test runner to use for regression tests on an ongoing basis once the runner is configured let s put it to work by defining a core smoke test for releases we should ensure this can run against our local test env first but i want to immediately start using this against some hosted staging demo environments in the short term so let s ensure that any urls are easy to configure and override e g will quickly become something like in addition let s signup for cypress cloud and publish results there it s a great product acceptance criteria must implement a new cypress runner to run tests targeting the local manual test environment see must support targeting different hosts via environment variables so we can run against staging demo etc must have a nox session e g nox s test that runs the cypress suite in headless mode must configure the test reporter to publish results to our cypress cloud account must write a single cypress test that runs a simple smoke test covering confirm can login to the admin ui confirm that the postgres mongo connectors are configured confirm that the privacy center can be accessed submit an access request via the privacy center approve the access request via the admin ui confirm the access request succeeds | 1 |
109,817 | 9,415,625,911 | IssuesEvent | 2019-04-10 13:04:24 | italia/spid | https://api.github.com/repos/italia/spid | closed | Richiesta di validazione dei metadati - Comune di Mapello | metadata nuovo md test | Buongiorno,
per conto del comune si richiede la verifica e il deploy dei metadati esposti all'URL:
https://sportellotelematico.comune.mapello.bg.it/simplesaml/module.php/saml/sp/metadata.php/spid
| 1.0 | Richiesta di validazione dei metadati - Comune di Mapello - Buongiorno,
per conto del comune si richiede la verifica e il deploy dei metadati esposti all'URL:
https://sportellotelematico.comune.mapello.bg.it/simplesaml/module.php/saml/sp/metadata.php/spid
| test | richiesta di validazione dei metadati comune di mapello buongiorno per conto del comune si richiede la verifica e il deploy dei metadati esposti all url | 1 |
293,603 | 25,309,074,523 | IssuesEvent | 2022-11-17 16:06:35 | Becksteinlab/MDPOW | https://api.github.com/repos/Becksteinlab/MDPOW | closed | DihedralAnalysis test_results_recursive2 fails most of the time | testing | The test for two dihedrals in DihedralAnalysis https://github.com/Becksteinlab/MDPOW/blob/2092cfd3f1cb006b38d01ec722dd2040f221a5c4/mdpow/tests/test_dihedral.py#L66-L83 started failing a few months ago with output like
```
_____________________ TestDihedral.test_results_recursive2 _____________________
self = <mdpow.tests.test_dihedral.TestDihedral object at 0x7f0585e22070>
def test_results_recursive2(self):
dh1 = self.Ens.select_atoms('name C11 or name C10 or name C9 or name C4')
dh2 = self.Ens.select_atoms('name C8 or name C4 or name C9 or name C10')
dh_run = DihedralAnalysis([dh1, dh2]).run(start=0, stop=4, step=1)
dh1_result = dh_run.results.loc[dh_run.results['selection'] == 'C4-C9-C10-C11']['dihedral']
dh2_result = dh_run.results.loc[dh_run.results['selection'] == 'C4-C8-C9-C10']['dihedral']
dh1_mean = circmean(dh1_result, high=180, low=-180)
dh2_mean = circmean(dh2_result, high=180, low=-180)
dh1_var = circvar(dh1_result, high=180, low=-180)
dh2_var = circvar(dh2_result, high=180, low=-180)
assert_almost_equal(self.DG48910_mean, dh1_mean, 6)
> assert_almost_equal(self.DG48910_var, dh1_var, 6)
E AssertionError:
E Arrays are not almost equal to 6 decimals
E ACTUAL: 1490.6576365537262
E DESIRED: 0.20311120667628568
```
See, for instance, https://github.com/Becksteinlab/MDPOW/runs/7905279813?check_suite_focus=true#step:10:160 and https://github.com/Becksteinlab/MDPOW/pull/213#issuecomment-1219889216 .
EDIT: ~The fact that the ACTUAL value is ~1490 is ludicrous for a dihedral angle and _possibly_ suggests a problem with circvar because in principle it should have remained in the range -180 < x < +180.~ Looking at the code, the *reference value* is 1490.6576365537262. That's still suspicious and should be looked into. The docs for [scipy.stats.circvar()](https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.circvar.html) state that `0 ≤ circvar ≤ 1`, i.e., it is not clear how we can get 1490.6576365537262 _and_ pass the test!
It does not always fail for all CI runners, only some of them. There's no clear pattern with older/newer Python versions that I can discern. Locally (on macOS Python 3.8) the test always passes even while 3.8 fails in the CI.
1. We need to understand if this is a bug in the DihedralAnalysis code or in the testing.
2. We need to make the tests work reliably again.
| 1.0 | DihedralAnalysis test_results_recursive2 fails most of the time - The test for two dihedrals in DihedralAnalysis https://github.com/Becksteinlab/MDPOW/blob/2092cfd3f1cb006b38d01ec722dd2040f221a5c4/mdpow/tests/test_dihedral.py#L66-L83 started failing a few months ago with output like
```
_____________________ TestDihedral.test_results_recursive2 _____________________
self = <mdpow.tests.test_dihedral.TestDihedral object at 0x7f0585e22070>
def test_results_recursive2(self):
dh1 = self.Ens.select_atoms('name C11 or name C10 or name C9 or name C4')
dh2 = self.Ens.select_atoms('name C8 or name C4 or name C9 or name C10')
dh_run = DihedralAnalysis([dh1, dh2]).run(start=0, stop=4, step=1)
dh1_result = dh_run.results.loc[dh_run.results['selection'] == 'C4-C9-C10-C11']['dihedral']
dh2_result = dh_run.results.loc[dh_run.results['selection'] == 'C4-C8-C9-C10']['dihedral']
dh1_mean = circmean(dh1_result, high=180, low=-180)
dh2_mean = circmean(dh2_result, high=180, low=-180)
dh1_var = circvar(dh1_result, high=180, low=-180)
dh2_var = circvar(dh2_result, high=180, low=-180)
assert_almost_equal(self.DG48910_mean, dh1_mean, 6)
> assert_almost_equal(self.DG48910_var, dh1_var, 6)
E AssertionError:
E Arrays are not almost equal to 6 decimals
E ACTUAL: 1490.6576365537262
E DESIRED: 0.20311120667628568
```
See, for instance, https://github.com/Becksteinlab/MDPOW/runs/7905279813?check_suite_focus=true#step:10:160 and https://github.com/Becksteinlab/MDPOW/pull/213#issuecomment-1219889216 .
EDIT: ~The fact that the ACTUAL value is ~1490 is ludicrous for a dihedral angle and _possibly_ suggests a problem with circvar because in principle it should have remained in the range -180 < x < +180.~ Looking at the code, the *reference value* is 1490.6576365537262. That's still suspicious and should be looked into. The docs for [scipy.stats.circvar()](https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.circvar.html) state that `0 ≤ circvar ≤ 1`, i.e., it is not clear how we can get 1490.6576365537262 _and_ pass the test!
It does not always fail for all CI runners, only some of them. There's no clear pattern with older/newer Python versions that I can discern. Locally (on macOS Python 3.8) the test always passes even while 3.8 fails in the CI.
1. We need to understand if this is a bug in the DihedralAnalysis code or in the testing.
2. We need to make the tests work reliably again.
| test | dihedralanalysis test results fails most of the time the test for two dihedrals in dihedralanalysis started failing a few months ago with output like testdihedral test results self def test results self self ens select atoms name or name or name or name self ens select atoms name or name or name or name dh run dihedralanalysis run start stop step result dh run results loc result dh run results loc mean circmean result high low mean circmean result high low var circvar result high low var circvar result high low assert almost equal self mean mean assert almost equal self var var e assertionerror e arrays are not almost equal to decimals e actual e desired see for instance and edit the fact that the actual value is is ludicrous for a dihedral angle and possibly suggests a problem with circvar because in principle it should have remained in the range x looking at the code the reference value is that s still suspicious and should be looked into the docs for state that ≤ circvar ≤ i e it is not clear how we can get and pass the test it does not always fail for all ci runners only some of them there s no clear pattern with older newer python versions that i can discern locally on macos python the test always passes even while fails in the ci we need to understand if this is a bug in the dihedralanalysis code or in the testing we need to make the tests work reliably again | 1 |
193,238 | 14,643,966,752 | IssuesEvent | 2020-12-25 19:59:04 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | kubernetes/kubernetes: pkg/controller/endpointslice/reconciler_test.go; 45 LoC | experts: :+1: fresh small test |
Found a possible issue in [kubernetes/kubernetes](https://www.github.com/kubernetes/kubernetes) at [pkg/controller/endpointslice/reconciler_test.go](https://github.com/kubernetes/kubernetes/blob/0a839c6c3b9489716e216b4fc9dd2f9c17948c16/pkg/controller/endpointslice/reconciler_test.go#L444-L488)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to slice at line 476 may start a goroutine
[Click here to see the code in its original context.](https://github.com/kubernetes/kubernetes/blob/0a839c6c3b9489716e216b4fc9dd2f9c17948c16/pkg/controller/endpointslice/reconciler_test.go#L444-L488)
<details>
<summary>Click here to show the 45 line(s) of Go which triggered the analyzer.</summary>
```go
for _, slice := range slices {
if !strings.HasPrefix(slice.Name, testCase.service.Name) {
t.Fatalf("Expected EndpointSlice name to start with %s, got %s", testCase.service.Name, slice.Name)
}
if !reflect.DeepEqual(testCase.expectedLabels, slice.Labels) {
t.Errorf("Expected EndpointSlice to have labels: %v , got %v", testCase.expectedLabels, slice.Labels)
}
if slice.Labels[discovery.LabelServiceName] != testCase.service.Name {
t.Fatalf("Expected EndpointSlice to have label set with %s value, got %s", testCase.service.Name, slice.Labels[discovery.LabelServiceName])
}
if slice.Annotations[corev1.EndpointsLastChangeTriggerTime] != triggerTime.Format(time.RFC3339Nano) {
t.Fatalf("Expected EndpointSlice trigger time annotation to be %s, got %s", triggerTime.Format(time.RFC3339Nano), slice.Annotations[corev1.EndpointsLastChangeTriggerTime])
}
// validate that this slice has address type matching expected
expectedEndPointList := testCase.expectedEndpointPerSlice[slice.AddressType]
if expectedEndPointList == nil {
t.Fatalf("address type %v is not expected", slice.AddressType)
}
if len(slice.Endpoints) != len(expectedEndPointList) {
t.Fatalf("Expected %v Endpoint, got %d", len(expectedEndPointList), len(slice.Endpoints))
}
// test is limited to *ONE* endpoint
endpoint := slice.Endpoints[0]
if !reflect.DeepEqual(endpoint, expectedEndPointList[0]) {
t.Fatalf("Expected endpoint: %+v, got: %+v", expectedEndPointList[0], endpoint)
}
expectTrackedResourceVersion(t, r.endpointSliceTracker, &slice, "100")
expectMetrics(t,
expectedMetrics{
desiredSlices: 1,
actualSlices: 1,
desiredEndpoints: 1,
addedPerSync: len(testCase.expectedEndpointPerSlice),
removedPerSync: 0,
numCreated: len(testCase.expectedEndpointPerSlice),
numUpdated: 0,
numDeleted: 0})
}
```
</details>
<details>
<summary>Click here to show extra information the analyzer produced.</summary>
```
The following dot graph describes paths through the callgraph that could lead to a function calling a goroutine:
digraph G {
"(writeback, 1)" -> {"(merge, 1)";}
"(scheduleBodyWrite, 0)" -> {}
"(shutDownIn, 1)" -> {}
"(sendPings, 1)" -> {}
"(WithTimeout, 2)" -> {"(WithDeadline, 2)";}
"(Example, 0)" -> {}
"(RpcError_ErrorCode, 1)" -> {"(CancelRequest, 1)";}
"(RetrieveOne, 4)" -> {"(sendConsumeCustomMetric, 5)";"(ServerPreferredResources, 1)";"(RediscoverNode, 1)";"(Run, 4)";"(sendLoop, 0)";"(SetUpAt, 2)";"(newConn, 2)";"(createMergedSnapshotMessage, 4)";"(createAndInitKubelet, 30)";"(NewClientConn, 3)";"(Evaluate, 1)";"(SpecDidComplete, 1)";"(ListenAndServeReadOnly, 3)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(on100, 0)";"(RestartNode, 1)";"(mountServiceAccountToken, 2)";"(NewSession, 2)";"(HandlePodReconcile, 1)";"(DeleteCollection, 4)";"(DiscoverNode, 1)";"(sendConsumeMemRequests, 4)";"(ConfigureTransports, 1)";"(BeforeSuiteDidRun, 1)";"(rollbackEtcd3MinorVersion, 2)";"(newETCD3Storage, 2)";"(keepAliveLoop, 2)";"(SpecSuiteWillBegin, 2)";"(SpecSuiteDidEnd, 1)";"(apply, 3)";"(Dispatch, 4)";"(handleResponse, 2)";"(recvLoop, 0)";"(Build, 3)";"(processPodOnNode, 5)";"(Sink, 0)";"(NewDelayingQueueWithCustomClock, 2)";"(RoundTripOpt, 2)";"(CloseNotify, 0)";"(startCompactorOnce, 2)";"(GetCollector, 1)";"(Chan, 0)";"(addPeerToProber, 6)";"(NewClientConnectionWithPings, 2)";"(PortForward, 3)";"(createVolume, 2)";"(NewProxier, 9)";"(NewStreamWatcher, 2)";"(Eventf, 7)";"(listenerRoutine, 0)";"(VerifySignature, 2)";"(UpdatePod, 1)";"(NewMainKubelet, 30)";"(sendMergedSnap, 1)";"(receive, 0)";"(StartTestServer, 4)";"(dispatch, 0)";"(HandleStreams, 2)";"(mergeService, 1)";"(Dispatch, 3)";"(runSTM, 2)";"(StartGarbageCollection, 0)";"(sendUpdates, 0)";"(roundTrip, 1)";"(UpdateTransport, 4)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(updateTransport, 5)";"(markDiskAsDeletable, 2)";"(applySnapshot, 2)";"(enable, 0)";"(recordToSink, 2)";"(newLessor, 3)";"(HandlePodCleanups, 0)";"(RunOnce, 1)";"(PortForward, 4)";"(getMetricsFromNode, 2)";"(NewAggregator, 4)";"(NewProxier, 13)";"(newClientConn, 2)";"(runOnce, 2)";"(AttachVolume, 2)";"(dialStreamLocal, 1)";"(AddPod, 1)";"(newSnapshotReaderCloser, 2)";"(sendConsumeCPURequests, 4)";"(AfterSuiteDidRun, 1)";"(newFilePipeline, 3)";"(startFrameWrite, 1)";"(BuildAuthn, 2)";"(newMux, 1)";"(dgemmParallel, 12)";"(SendSnapshot, 1)";"(monitorSize, 2)";"(podsRunning, 2)";"(GenerateRegisterPluginFunc, 4)";"(applyEntries, 2)";"(newWatcherGrpcStream, 1)";"(HandlePodUpdates, 1)";"(CreatePD, 1)";"(verify, 1)";"(newDecoratedWatcher, 2)";"(handleConnection, 2)";"(Launch, 1)";"(scheduleFrameWrite, 0)";"(handleSignal, 1)";"(removeAndReAdd, 1)";"(callK, 1)";"(newProxyServer, 3)";"(ListenAndServe, 8)";"(SyncServices, 1)";"(StartInterceptingOutput, 0)";"(createCall, 5)";"(CreateVolume, 3)";"(newHTTP2Transport, 2)";"(Verify, 2)";"(startKubelet, 6)";"(StartNode, 2)";"(startPeer, 4)";"(waitCancelSubstreams, 1)";"(initializeCSINode, 1)";"(startDBSizeMonitorPerEndpoint, 2)";"(clientHandshake, 2)";}
"(UpdateObjectCount, 2)" -> {"(newTimeBudget, 1)";}
"(AddHTTP, 3)" -> {}
"(Search, 2)" -> {"(Execute, 2)";}
"(FromContext, 1)" -> {"(Unknown, 1)";}
"(callTRFunc, 3)" -> {}
"(replyToClient, 2)" -> {"(deadlineLoop, 0)";}
"(runOnce, 2)" -> {"(getFailedContainers, 1)";}
"(startCaching, 1)" -> {}
"(convertProcessError, 2)" -> {"(processSignalResult, 2)";}
"(processData, 0)" -> {"(onEachWrite, 0)";}
"(syncVictimsLoop, 0)" -> {"(moveVictims, 0)";}
"(cleanupOrphanedPodCgroups, 3)" -> {}
"(InstallAPIs, 3)" -> {"(InstallAPIGroups, 1)";}
"(NewTicker, 1)" -> {"(yaml_emitter_state_machine, 2)";}
"(newClientV3, 1)" -> {"(toGRPC, 1)";}
"(pathsToRemove, 2)" -> {"(Difference, 1)";}
"(LastIndex, 0)" -> {"(passesVerifyFilter, 2)";"(initializeCSINode, 1)";"(dispatch, 0)";"(runSTM, 2)";"(freepages, 0)";"(newETCD3Storage, 2)";"(NewClientConnectionWithPings, 2)";"(roundTrip, 1)";"(NormalizedName, 1)";"(AddPod, 1)";"(RoundTripOpt, 2)";}
"(TokenVolumeSource, 0)" -> {}
"(newAffinityTerm, 2)" -> {"(GetNamespacesFromPodAffinityTerm, 2)";}
"(waitDelete, 4)" -> {"(Watch, 3)";"(freepages, 0)";}
"(wrapErr, 1)" -> {}
"(NewMainKubelet, 30)" -> {"(defaultNodeStatusFuncs, 0)";}
"(serveUDP, 1)" -> {"(spawnWorker, 1)";}
"(string, 1)" -> {"(DecodeRuneInString, 1)";}
"(Encode, 2)" -> {"(msgHeader, 2)";}
"(NewClusterV3, 2)" -> {"(NewClientV3, 1)";}
"(legacyLogSymlink, 4)" -> {"(logSymlink, 4)";}
"(getAgentPoolScaleSets, 1)" -> {"(CreateOrUpdateVMSS, 3)";}
"(msgHeader, 2)" -> {"(Do, 2)";"(wrapErr, 1)";}
"(handleConnection, 2)" -> {}
"(Exec, 2)" -> {"(AbsHostPath, 1)";}
"(RemoveImage, 1)" -> {"(RemoveImage, 2)";}
"(standardMethodArgumentDecode, 4)" -> {}
"(logSymlink, 4)" -> {"(transformStartContainerError, 1)";"(CreateComputeSystem, 3)";}
"(SpecSuiteDidEnd, 1)" -> {"(AnnounceSpecRunCompletion, 2)";}
"(From, 1)" -> {"(To, 1)";}
"(ToJSONStream, 1)" -> {"(Chan, 0)";"(on100, 0)";"(VerifySignature, 2)";"(serve, 0)";"(lockedUpdateLoadBalancerHosts, 2)";"(startFrameWrite, 1)";"(CloseNotify, 0)";"(createVolume, 2)";"(NewTimeoutListener, 5)";"(Eventf, 7)";"(newSnapshotReaderCloser, 2)";"(PortForward, 3)";"(SendSnapshot, 1)";"(Dispatch, 4)";}
"(OrNil, 0)" -> {}
"(add, 1)" -> {"(evict, 0)";}
"(Version, 0)" -> {"(NewVersion, 1)";}
"(, 2)" -> {"(getRequest, 1)";}
"(CreateHealthCheck, 1)" -> {"(HealthChecks, 0)";"(newETCD3HealthCheck, 1)";}
"(toGRPC, 1)" -> {"(ElectionServerToElectionClient, 1)";}
"(GRPCStatus, 0)" -> {"(Exp, 1)";"(SetTypeForKey, 2)";"(uint64, 0)";"(getFileType, 1)";}
"(erorr, 0)" -> {"(processSignalResult, 2)";}
"(startPeer, 4)" -> {"(newSnapshotSender, 4)";}
"(defaultNodeStatusFuncs, 0)" -> {"(VolumesInUse, 2)";}
"(NewProxier, 9)" -> {"(NewCustomProxier, 10)";}
"(layerPath, 2)" -> {}
"(NewOpenIDMetadata, 4)" -> {"(openIDKeysetJSON, 1)";}
"(Sync, 0)" -> {"(errnoErr, 1)";}
"(newCluster, 2)" -> {"(fillClusterForMembers, 0)";}
"(IterateUsing, 2)" -> {"(eachMapEntry, 2)";}
"(removeBlock, 1)" -> {"(Clear, 0)";"(paragraph, 1)";}
"(Clone, 0)" -> {"(NewTLS, 1)";}
"(Evaluate, 1)" -> {}
"(waitForChanges, 0)" -> {"(ChangeEvents, 2)";}
"(Sign, 0)" -> {"(Run, 1)";}
"(Head, 2)" -> {"(keyIndex, 1)";}
"(configureHTTP2Transport, 1)" -> {"(pingTimeoutSeconds, 0)";}
"(Config, 0)" -> {"(toStringSlice, 1)";}
"(New, 3)" -> {"(processFlag, 1)";}
"(hcsCloseProcess, 1)" -> {}
"(SortStrings, 1)" -> {}
"(topIndex, 0)" -> {"(Dispatch, 4)";}
"(max, 1)" -> {}
"(updateResolverState, 2)" -> {"(updateClientConnState, 1)";}
"(numLineBreaks, 1)" -> {"(newWatcherGrpcStream, 1)";}
"(serveListen, 0)" -> {"(serveConn, 1)";}
"(monitorSize, 2)" -> {"(monitorResizeEvents, 3)";}
"(empty, 0)" -> {"(StringAffineComparable, 1)";}
"(dispatch, 0)" -> {}
"(Eventf, 7)" -> {}
"(handleReadLoopResponse, 1)" -> {}
"(append, 1)" -> {}
"(flushColdBuf, 0)" -> {"(maybeRotateStreams, 0)";}
"(sendLoop, 0)" -> {"(RequestProgress, 1)";}
"(OpenComputeSystem, 2)" -> {}
"(GetClientConn, 2)" -> {"(getClientConn, 3)";}
"(TunnelSyncHealthChecker, 1)" -> {"(SecondsSinceSSHKeySync, 0)";}
"(Done, 0)" -> {"(erorr, 0)";}
"(CloseNotify, 0)" -> {}
"(switchBalancer, 1)" -> {"(newCCBalancerWrapper, 3)";}
"(getFailedContainers, 1)" -> {"(syncPod, 1)";}
"(processFlag, 1)" -> {"(setState, 1)";}
"(ErrAllFailed, 1)" -> {"(verify, 1)";}
"(StartGatheringData, 0)" -> {}
"(RemoveImage, 2)" -> {"(ImageRemove, 3)";}
"(Zcopy, 5)" -> {"(BrowseDiagnosticLog, 3)";}
"(hive, 1)" -> {"(StartGarbageCollection, 0)";"(sendUpdates, 0)";"(NewProxier, 9)";"(ListenAndServe, 8)";"(ListenAndServeReadOnly, 3)";"(NewMainKubelet, 30)";"(runOnce, 2)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(StartInterceptingOutput, 0)";}
"(cancel, 2)" -> {"(removeChild, 2)";}
"(NewCacherFromConfig, 1)" -> {"(startCaching, 1)";}
"(Keys, 1)" -> {"(KeysPrefix, 2)";}
"(freeTrace, 1)" -> {"(provider, 0)";}
"(resetTransport, 0)" -> {"(startHealthCheck, 1)";}
"(nodev, 1)" -> {"(node, 2)";}
"(Launch, 1)" -> {"(waitVersion, 0)";}
"(walkParameters, 1)" -> {"(HandleStreams, 2)";"(StartRequest, 7)";"(resolveNow, 1)";"(NewServerTransport, 3)";"(applyServiceConfigAndBalancer, 2)";"(NewClusterV3, 2)";"(makeConnectedServerPipe, 0)";}
"(DefaultCollector, 1)" -> {"(newConn, 2)";"(dialStreamLocal, 1)";"(createCall, 5)";}
"(startStreamWriter, 6)" -> {}
"(GetContainerInfoV2, 2)" -> {"(OrNil, 0)";}
"(GetRequestedContainersInfo, 2)" -> {"(OrNil, 0)";}
"(effectivelyEquals, 1)" -> {"(IsSelected, 1)";}
"(PollImmediateUntil, 3)" -> {"(PollUntil, 3)";}
"(Remove, 1)" -> {"(removeBlock, 1)";}
"(errorf, 3)" -> {"(CreateHealthCheck, 1)";"(executeCredentialProcess, 0)";"(createCall, 5)";"(ConfigureTransports, 1)";"(RediscoverNode, 1)";"(createDialer, 0)";}
"(executeCredentialProcess, 0)" -> {}
"(reflectValue, 1)" -> {"(valueEncoder, 1)";}
"(GetLatestPage, 0)" -> {"(RawSyscallNoError, 4)";}
"(SendRequest, 3)" -> {"(fn, 1)";}
"(Get, 1)" -> {"(GetUsing, 2)";}
"(createAndInitKubelet, 30)" -> {"(StartGarbageCollection, 0)";}
"(NewTLS, 1)" -> {"(AddHTTP, 3)";"(mustNewMember, 1)";"(Launch, 0)";"(startStreamWriter, 6)";}
"(deleteContainersInPod, 3)" -> {"(getContainersToDeleteInPod, 3)";}
"(NewClientHandler, 3)" -> {"(requestLogger, 2)";}
"(CreateOrUpdateVMSS, 3)" -> {"(PutResources, 2)";}
"(update, 4)" -> {"(CanSet, 0)";}
"(yaml_parser_set_input_string, 2)" -> {}
"(debugResponse, 1)" -> {"(newTeeReader, 2)";}
"(mkSTM, 2)" -> {"(runSTM, 2)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(GetClientConn, 2)";}
"(write, 0)" -> {"(fdatasync, 1)";}
"(newFilePipeline, 3)" -> {}
"(newCCBalancerWrapper, 3)" -> {"(Build, 2)";}
"(handler, 0)" -> {"(Lstat, 1)";}
"(formatUnequalValues, 2)" -> {"(truncatingFormat, 1)";}
"(V, 1)" -> {"(setV, 1)";}
"(GetLengthSid, 1)" -> {"(GetFileType, 1)";}
"(convertSliceNumbers, 2)" -> {"(convertNumber, 1)";}
"(runHandler, 1)" -> {"(handler, 0)";}
"(Do, 1)" -> {"(ReadAll, 1)";}
"(parseProposeCtxErr, 2)" -> {"(Upload, 4)";}
"(GenerateBulkVolumeVerifyFunc, 4)" -> {"(OperationCompleteHook, 2)";"(GetFullQualifiedPluginNameForVolume, 2)";}
"(shutdown, 1)" -> {}
"(copyStderr, 1)" -> {}
"(getFloatingIPByPortID, 2)" -> {"(selectLoadBalancer, 4)";}
"(Set, 1)" -> {"(isLiteral, 1)";}
"(checkTxnAuth, 3)" -> {"(checkTxnReqsPermission, 3)";}
"(StartTestServerOrDie, 4)" -> {"(StartTestServer, 4)";}
"(SetEndpoints, 1)" -> {"(NewAddress, 1)";}
"(kind, 0)" -> {"(uint64, 0)";}
"(ok, 0)" -> {"(satisfiedExpectations, 0)";}
"(recvLoop, 0)" -> {}
"(isControl, 0)" -> {}
"(Add, 2)" -> {"(SetScale, 1)";}
"(Signal, 1)" -> {"(Marshal, 1)";}
"(DeleteProducer, 1)" -> {}
"(Next, 0)" -> {"(lookup, 0)";}
"(NsHandle, 1)" -> {"(callopen64, 3)";}
"(HealthChecks, 0)" -> {}
"(listenerRoutine, 0)" -> {"(makeConnectedServerPipe, 0)";}
"(Do, 0)" -> {}
"(printDepth, 5)" -> {"(Filter, 1)";"(output, 6)";"(Bytes, 0)";"(getBuffer, 0)";}
"(sendConsumeMemRequests, 4)" -> {}
"(getRequestedContainers, 2)" -> {}
"(executePreStopHook, 4)" -> {}
"(Properties, 4)" -> {"(DefaultCollector, 1)";}
"(readAll, 2)" -> {"(SetUnknown, 1)";}
"(autoSync, 0)" -> {"(Lvl, 1)";}
"(Get, 4)" -> {"(NewError, 3)";}
"(Insert, 1)" -> {"(Descend, 1)";}
"(AddPod, 1)" -> {"(VisitPodConfigmapNames, 2)";}
"(newBpsLoop, 1)" -> {}
"(newBackend, 1)" -> {"(run, 0)";}
"(AddExtensionType, 2)" -> {"(getTrimmed, 1)";"(NewTicker, 1)";"(yaml_emitter_emit, 2)";}
"(Unlock, 1)" -> {"(Delete, 2)";}
"(getWrappedWorkerFunc, 1)" -> {}
"(NewKey, 2)" -> {}
"(createMetadatas, 1)" -> {"(WithTTL, 1)";}
"(AppendUint64, 1)" -> {"(AppendUint, 1)";}
"(apply, 3)" -> {"(AddNamedImport, 4)";}
"(GetCgroupStats, 2)" -> {"(cadvisorInfoToNetworkStats, 1)";}
"(rollbackEtcd3MinorVersion, 2)" -> {"(Snapshot, 2)";}
"(newTLSListener, 3)" -> {"(acceptLoop, 0)";}
"(Transform, 3)" -> {"(kind, 0)";}
"(maybeRotateStreams, 0)" -> {"(Descend, 1)";}
"(NewConnection, 2)" -> {"(monitor, 0)";}
"(previousDestroy, 0)" -> {"(UpdateObjectCount, 2)";}
"(waitCancelSubstreams, 1)" -> {}
"(listPodStats, 1)" -> {"(makePodStorageStats, 2)";}
"(output, 6)" -> {"(timeoutFlush, 1)";"(int64, 1)";"(Format, 1)";"(Reset, 0)";"(Error, 0)";}
"(Dispatch, 3)" -> {"(Dispatch, 4)";}
"(NewClientV3, 1)" -> {"(newClientV3, 1)";}
"(provider, 0)" -> {"(newWatcherGrpcStream, 1)";"(Dispatch, 4)";}
"(ListPodStats, 0)" -> {"(listPodStats, 1)";}
"(DialWithConn, 4)" -> {"(newMux, 1)";"(HandleStreams, 2)";"(AttachVolume, 2)";"(Chan, 0)";"(removeAndReAdd, 1)";"(DeleteCollection, 4)";"(startFrameWrite, 1)";"(Dispatch, 4)";"(NewMainKubelet, 30)";"(sendUpdates, 0)";"(rollbackEtcd3MinorVersion, 2)";"(HandlePodCleanups, 0)";"(GenerateRegisterPluginFunc, 4)";"(runSTM, 2)";"(SyncServices, 1)";"(StartInterceptingOutput, 0)";"(StartTestServer, 4)";"(CloseNotify, 0)";"(sendConsumeCustomMetric, 5)";"(CreatePD, 1)";"(startDBSizeMonitorPerEndpoint, 2)";"(Launch, 0)";"(mergeService, 1)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(dispatch, 0)";"(sendConsumeMemRequests, 4)";"(NewProxier, 9)";"(handleConnection, 2)";"(CreateVolume, 3)";"(newSnapshotReaderCloser, 2)";"(dgemmParallel, 12)";"(updateTransport, 5)";"(NewDelayingQueueWithCustomClock, 2)";"(ListenAndServe, 8)";"(sendConsumeCPURequests, 4)";"(ListenAndServeReadOnly, 3)";"(Eventf, 7)";"(PortForward, 3)";"(createVolume, 2)";"(VerifySignature, 2)";"(NewClientConnectionWithPings, 2)";"(startCompactorOnce, 2)";"(SendSnapshot, 1)";"(processPodOnNode, 5)";"(newWatcherGrpcStream, 1)";"(recordToSink, 2)";"(Run, 4)";"(UpdatePod, 1)";"(on100, 0)";"(monitorResizeEvents, 3)";"(DiscoverNode, 1)";"(clientHandshake, 2)";"(StartGarbageCollection, 0)";"(runOnce, 2)";"(handleSignal, 1)";"(verify, 1)";}
"(fetchGroupVersionResources, 2)" -> {}
"(eachMapEntry, 2)" -> {"(dialStreamLocal, 1)";"(Launch, 1)";"(Build, 3)";"(createAndInitKubelet, 30)";"(UpdateTransport, 4)";"(newFilePipeline, 3)";"(newSession, 2)";"(BuildAuthn, 2)";"(StartNode, 2)";"(StartTestServerOrDie, 4)";"(callK, 1)";"(MigrateIfNeeded, 1)";"(inWorker, 0)";"(Sink, 0)";"(receive, 0)";"(addPeerToProber, 6)";"(mountServiceAccountToken, 2)";"(newConn, 2)";"(NewProxier, 13)";"(SetUpAt, 2)";"(RediscoverNode, 1)";"(Verify, 2)";"(copyStdin, 0)";"(StartGatheringData, 0)";"(enable, 0)";"(ServerGroupsAndResources, 1)";"(NewAggregator, 4)";"(waitForAvailableEndpoint, 1)";"(newClientConn, 2)";"(roundTrip, 1)";"(handleResizes, 0)";"(Evaluate, 1)";"(sendLoop, 0)";"(RoundTripOpt, 2)";"(Dispatch, 3)";"(createCall, 5)";"(freepages, 0)";"(copyStderr, 1)";"(initializeCSINode, 1)";"(CopyAllLogs, 4)";"(newETCD3Storage, 2)";"(NewClientConn, 3)";"(Dgemm, 13)";"(applySnapshot, 2)";"(copyStdout, 1)";"(createMergedSnapshotMessage, 4)";"(scheduleFrameWrite, 0)";"(passesVerifyFilter, 2)";"(ListenForSignal, 1)";"(recvLoop, 0)";"(NewSession, 2)";"(RunOnce, 1)";"(listenerRoutine, 0)";"(newHTTP2Transport, 2)";"(newProxyServer, 3)";"(watchErrorStream, 2)";"(sendMergedSnap, 1)";"(podsRunning, 2)";"(handleResponse, 2)";"(SpecSuiteWillBegin, 2)";"(SpecSuiteDidEnd, 1)";"(PortForward, 4)";"(waitCancelSubstreams, 1)";"(applyEntries, 2)";"(startKubelet, 6)";"(startPeer, 4)";"(createPD, 1)";"(RestartNode, 1)";"(newDecoratedWatcher, 2)";"(AddPod, 1)";"(getMetricsFromNode, 2)";"(WatchPods, 4)";"(GetCollector, 1)";"(NewStreamWatcher, 2)";"(monitorSize, 2)";"(DefineTests, 2)";"(newLessor, 3)";}
"(CreateComputeSystem, 3)" -> {}
"(Values, 0)" -> {"(retrieveUnexportedField, 3)";}
"(yaml_emitter_emit, 2)" -> {"(handleInterrupt, 0)";"(ConfigureTransports, 1)";"(callK, 1)";}
"(newTeeReader, 2)" -> {"(newWatcherGrpcStream, 1)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(Connect, 0)";"(GetClientConn, 2)";}
"(Flush, 0)" -> {"(Split, 2)";"(lockAndFlushAll, 0)";}
"(Split, 2)" -> {"(writeDataFromHandler, 3)";}
"(PortForward, 4)" -> {"(PortForward, 3)";}
"(insert, 2)" -> {}
"(NewNamingConditionController, 2)" -> {"(AddEventHandler, 1)";}
"(sweetenFields, 1)" -> {"(append, 1)";"(Remove, 1)";"(pathsToRemove, 2)";"(Wait, 0)";"(openExistingOrNew, 1)";"(Header, 0)";"(Decode, 1)";"(Do, 1)";"(V, 1)";"(Transform, 3)";"(Sync, 0)";"(Call, 5)";"(Flush, 0)";"(Index, 2)";"(flushColdBuf, 0)";"(New, 3)";"(Add, 2)";"(Parse, 1)";"(New, 0)";"(Update, 3)";"(Add, 1)";"(Send, 1)";"(Put, 1)";"(Copy, 2)";"(rotate, 0)";"(Encode, 2)";"(Reset, 0)";}
"(NewController, 1)" -> {"(newCallSet, 0)";}
"(Client, 1)" -> {"(NewClient, 2)";}
"(dedup, 1)" -> {"(LoadSources, 3)";}
"(Infof, 2)" -> {"(Printf, 2)";}
"(sendKeepAliveLoop, 1)" -> {"(LeaseKeepAlive, 1)";}
"(onEachWrite, 0)" -> {"(SendSnapshot, 1)";"(newSnapshotReaderCloser, 2)";"(CloseNotify, 0)";"(on100, 0)";"(startFrameWrite, 1)";"(PortForward, 3)";}
"(delete, 1)" -> {"(empty, 0)";}
"(testScriptInPod, 4)" -> {"(GetTestImage, 1)";}
"(GetUsing, 2)" -> {"(findJsonNameField, 1)";}
"(outPayload, 5)" -> {"(finishStream, 5)";"(sendConsumeCPURequests, 4)";"(CloseNotify, 0)";"(DeleteCollection, 4)";"(sendConsumeCustomMetric, 5)";"(runSTM, 2)";"(GenerateBulkVolumeVerifyFunc, 4)";"(NewClientConnectionWithPings, 2)";"(Run, 4)";"(newWatcherGrpcStream, 1)";"(recordToSink, 2)";"(addEvent, 3)";"(NewDelayingQueueWithCustomClock, 2)";"(RoundTripOpt, 2)";"(dispatch, 0)";"(sendConsumeMemRequests, 4)";"(HandleStreams, 2)";"(roundTrip, 1)";"(markDiskAsDeletable, 2)";}
"(Quo, 2)" -> {"(GetClientConn, 2)";"(dispatch, 0)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(NewClientConnectionWithPings, 2)";}
"(StartNode, 2)" -> {}
"(NewSession, 2)" -> {"(KeepAlive, 2)";}
"(createVolume, 2)" -> {"(TokenVolumeSource, 0)";}
"(Load, 0)" -> {"(Start, 1)";"(Get, 1)";"(Next, 0)";"(PollImmediateUntil, 3)";"(NewDiscoveryController, 3)";"(stopServers, 2)";"(newBackend, 1)";"(sync, 0)";"(Encode, 2)";"(NewREST, 2)";"(commit, 1)";"(Add, 2)";"(NewWithTLSConfig, 2)";"(Read, 1)";"(SetEndpoints, 1)";"(Node, 0)";"(getDelegate, 0)";"(ListPodStatsAndUpdateCPUNanoCoreUsage, 0)";"(From, 1)";"(keyToFilename, 1)";"(Start, 0)";"(Unmarshal, 2)";"(HcsCloseProcess, 2)";"(Set, 1)";"(Handle, 2)";"(New, 3)";"(, 0)";"(Terminate, 0)";"(V1, 0)";"(Verifier, 1)";"(Get, 4)";"(Write, 1)";"(ImportLayer, 4)";"(Search, 2)";"(remove, 1)";"(WithTimeout, 2)";"(ListPodStats, 0)";"(NewNamingConditionController, 2)";"(New, 1)";"(Initialize, 2)";"(ServiceTicket, 3)";"(ok, 0)";"(NewCustomResourceDefinitionHandler, 15)";"(get, 1)";"(SendRequest, 3)";"(close, 0)";"(Lock, 1)";"(NewCRDFinalizer, 3)";"(shutdown, 1)";"(NewEstablishingController, 2)";"(reset, 0)";"(Config, 0)";"(newClient, 1)";"(Get, 3)";"(Get, 2)";"(delete, 1)";"(Sub, 1)";"(V, 1)";"(enableCSM, 3)";"(NewDesc, 4)";"(InstallAPIs, 3)";"(fetch, 1)";"(Reset, 0)";"(Zero, 0)";"(NewConditionController, 2)";"(tearDown, 1)";"(Wait, 0)";"(Decode, 1)";"(GetFromPath, 1)";"(Do, 0)";"(ToFloat64, 1)";"(Keys, 1)";"(unregisterCallback, 1)";"(NewClient, 2)";"(add, 1)";"(Do, 1)";"(Remove, 1)";"(Apiextensions, 0)";"(PollUntil, 3)";"(Run, 3)";"(New, 0)";"(Delete, 1)";"(Run, 2)";"(Run, 1)";"(apply, 2)";"(Add, 1)";"(newStore, 1)";"(Transform, 3)";"(Install, 1)";"(New, 2)";"(parse, 1)";"(NewClusterAuthenticationTrustController, 2)";"(New, 6)";"(NewSessionWithOptions, 1)";"(Sign, 0)";"(Equal, 3)";"(Sync, 0)";"(Flush, 0)";"(NewOpenIDMetadata, 4)";"(GetCgroupStats, 2)";"(reset, 1)";"(Get, 0)";"(Client, 1)";"(Parse, 1)";"(Put, 1)";"(NewCtxClient, 1)";"(Size, 0)";"(init, 0)";"(HcsCloseComputeSystem, 2)";"(Send, 1)";"(NewKubernetesAPIApprovalPolicyConformantConditionController, 2)";"(Format, 1)";"(newAuthenticator, 2)";"(GetWithContext, 1)";"(Unlock, 1)";"(Revoke, 2)";"(NewLeaseFromLeaseClient, 3)";"(Error, 0)";"(ToJSON, 1)";"(installTunneler, 2)";"(NewController, 1)";}
"(poller, 2)" -> {}
"(awaitRequestCancel, 1)" -> {}
"(NewCustomProxier, 10)" -> {"(createProxier, 10)";}
"(ClockWait, 3)" -> {}
"(NewDefaultTableConvertor, 1)" -> {"(previousDestroy, 0)";}
"(Index, 2)" -> {"(Insert, 1)";}
"(Untranspose, 0)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(newWatcherGrpcStream, 1)";"(DeleteCollection, 4)";"(sendConsumeCPURequests, 4)";"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";}
"(on100, 0)" -> {}
"(UserDefinedMetricType, 1)" -> {"(getRequestedContainers, 2)";}
"(GarbageCollect, 0)" -> {"(DrainedAt, 0)";}
"(ListenAndServeKubeletServer, 10)" -> {"(ListenAndServe, 0)";}
"(ListenAndServePodResources, 0)" -> {"(ListenAndServePodResources, 4)";}
"(printWithFileLine, 7)" -> {"(create, 3)";}
"(allocate, 0)" -> {"(roundTrip, 1)";"(GetRequestedContainersInfo, 2)";"(fetchContainerStats, 1)";"(RoundTripOpt, 2)";"(callTTBFunc, 3)";"(apply, 3)";"(GetContainerInfoV2, 2)";}
"(totalSecretBytes, 1)" -> {}
"(GenerateRegisterPluginFunc, 4)" -> {"(RegisterPlugin, 3)";}
"(containerStartSpec, 1)" -> {"(legacyLogSymlink, 4)";}
"(printDelimiter, 0)" -> {"(ChangeEvents, 2)";}
"(UpdateLoadBalancer, 4)" -> {"(updateSecurityGroup, 4)";}
"(Copy, 2)" -> {"(Zcopy, 5)";}
"(StartGarbageCollection, 0)" -> {"(GarbageCollect, 0)";}
"(spawnWorker, 1)" -> {}
"(stopServers, 2)" -> {"(GracefulStop, 0)";}
"(String, 0)" -> {"(Iterate, 1)";}
"(makePodStorageStats, 2)" -> {"(cadvisorInfoToAcceleratorStats, 1)";}
"(newDelayingQueue, 3)" -> {}
"(processPodOnNode, 5)" -> {"(GetWorkerUnsafe, 1)";}
"(platformDoesNotSupportError, 1)" -> {"(isFeatureInRange, 2)";}
"(FatalDepth, 2)" -> {"(printDepth, 5)";}
"(handleResponse, 2)" -> {"(awaitRequestCancel, 1)";}
"(getBuffer, 0)" -> {"(Reset, 0)";}
"(initializeCSINode, 1)" -> {}
"(wrapTLS, 3)" -> {"(newTLSListener, 3)";}
"(expectTrackedResourceVersion, 4)" -> {"(Fatalf, 2)";"(Errorf, 4)";}
"(merge, 1)" -> {"(WithLazyPrefix, 1)";}
"(handleResizes, 0)" -> {}
"(RequestProgress, 1)" -> {}
"(Properties, 2)" -> {}
"(errnoErr, 1)" -> {"(StrategicMerge, 1)";"(msync, 1)";}
"(getFileType, 1)" -> {}
"(moveVictims, 0)" -> {"(newDelayingQueue, 3)";"(UpdateLoadBalancer, 4)";"(GetLoadBalancer, 3)";}
"(setV, 1)" -> {"(FileLine, 1)";}
"(compact, 3)" -> {"(Launch, 0)";"(NewProxier, 9)";"(ListenAndServe, 8)";"(recordToSink, 2)";"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(runOnce, 2)";"(monitorResizeEvents, 3)";"(sendUpdates, 0)";"(StartInterceptingOutput, 0)";"(newHTTP2Transport, 2)";"(ListenAndServeReadOnly, 3)";}
"(SetUnknown, 1)" -> {"(NewClientConnectionWithPings, 2)";"(markDiskAsDeletable, 2)";"(dispatch, 0)";}
"(callopen64, 3)" -> {"(openDir, 1)";}
"(newCond, 0)" -> {"(method, 0)";}
"(format, 2)" -> {"(format, 3)";}
"(Caller, 1)" -> {"(Next, 0)";}
"(ParseNetlinkMessage, 1)" -> {}
"(Printf, 2)" -> {"(output, 1)";}
"(isMembersEqual, 2)" -> {"(SortableMemberSliceByPeerURLs, 1)";}
"(GetCollector, 1)" -> {"(newCollector, 2)";}
"(visitContainerConfigmapNames, 2)" -> {"(newAffinityTerm, 2)";"(newIntSet, 0)";}
"(installTunneler, 2)" -> {"(TunnelSyncHealthChecker, 1)";}
"(storeScratch, 4)" -> {"(inBounds, 3)";}
"(Add, 1)" -> {"(ref, 0)";}
"(ToFloat64E, 1)" -> {"(indirect, 1)";}
"(Descend, 1)" -> {"(iterate, 6)";}
"(Sink, 0)" -> {}
"(shouldRetryRequest, 3)" -> {"(canRetryError, 1)";}
"(Filter, 1)" -> {"(Set, 2)";}
"(SetThumbprint, 2)" -> {}
"(AddEventHandler, 1)" -> {"(AddEventHandlerWithResyncPeriod, 2)";}
"(connect, 0)" -> {"(resetTransport, 0)";}
"(CreateDisk, 5)" -> {}
"(listenGRPC, 0)" -> {"(newBridge, 1)";}
"(ImageRemove, 3)" -> {"(executePreStopHook, 4)";}
"(get, 0)" -> {"(Load, 0)";}
"(GetPath, 0)" -> {"(GetIntelRdtPath, 1)";}
"(reset, 1)" -> {"(newReadOnly, 1)";}
"(HandleStreams, 2)" -> {}
"(waitVersion, 0)" -> {"(isMembersEqual, 2)";}
"(startHealthCheck, 1)" -> {"(healthCheckFunc, 4)";}
"(Decode, 1)" -> {"(newDecoder, 0)";}
"(Observe, 1)" -> {"(asyncFlush, 1)";}
"(startKubelet, 6)" -> {"(ListenAndServePodResources, 0)";}
"(Upload, 4)" -> {"(NewReader, 4)";}
"(decodeArguments, 3)" -> {"(standardMethodArgumentDecode, 4)";}
"(, 0)" -> {"(newReflectCache, 0)";}
"(Line, 0)" -> {"(lookup, 0)";}
"(finishStream, 5)" -> {"(addEvent, 3)";}
"(NewKubernetesAPIApprovalPolicyConformantConditionController, 2)" -> {"(AddEventHandler, 1)";}
"(Wrapf, 2)" -> {"(unmarshal, 3)";}
"(NewREST, 2)" -> {"(NewDefaultTableConvertor, 1)";}
"(Error, 0)" -> {"(Quote, 1)";"(format, 4)";}
"(Dgemm, 13)" -> {"(dgemmParallel, 12)";}
"(OpenContainer, 1)" -> {"(OpenComputeSystem, 2)";}
"(releaseEntry, 1)" -> {"(Put, 1)";"(Lock, 0)";"(log, 2)";"(Get, 0)";"(Unlock, 0)";"(Format, 4)";}
"(ref, 0)" -> {"(freeTrace, 1)";}
"(rebuild, 2)" -> {"(ReplaceOrInsert, 1)";}
"(Exp, 1)" -> {}
"(handleSignal, 1)" -> {"(DeliverSignal, 3)";}
"(PollUntil, 3)" -> {"(poller, 2)";}
"(newSession, 3)" -> {}
"(mkV2Events, 1)" -> {"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(Upload, 4)";"(sendConsumeCPURequests, 4)";}
"(Lvl, 1)" -> {}
"(isUnderlinedHeading, 1)" -> {"(RoundTripOpt, 2)";"(Dispatch, 4)";"(addEvent, 3)";"(roundTrip, 1)";}
"(int64, 1)" -> {}
"(keyIndex, 1)" -> {"(mkV2Events, 1)";}
"(RunOnce, 1)" -> {"(runOnce, 2)";}
"(GenerateAttachVolumeFunc, 2)" -> {}
"(Finish, 0)" -> {"(Nanoseconds, 0)";}
"(NewCustomResourceDefinitionHandler, 15)" -> {"(Store, 1)";}
"(ProtoReflect, 0)" -> {"(Resolve, 1)";"(SetUnknown, 1)";}
"(RawVector, 0)" -> {"(Untranspose, 0)";}
"(PathPrefix, 0)" -> {"(NewClientV3, 1)";}
"(NewChecker, 4)" -> {}
"(registerEvent, 3)" -> {"(addEventFile, 5)";}
"(NewUnsafeSet, 1)" -> {"(Init, 1)";}
"(ConfigureTransports, 1)" -> {"(configureTransports, 1)";}
"(dgemmParallel, 12)" -> {}
"(reset, 2)" -> {"(peekType, 0)";}
"(formatHeader, 3)" -> {"(Next, 0)";"(Write, 1)";}
"(Strings, 1)" -> {"(Reset, 0)";}
"(makeConnectedServerPipe, 0)" -> {}
"(clientHandshake, 2)" -> {"(clientAuthenticate, 1)";}
"(authCapabilityHandler, 1)" -> {}
"(Unlock, 0)" -> {"(writeback, 1)";}
"(Do, 2)" -> {"(Delims, 2)";}
"(addEvent, 3)" -> {}
"(GetNamespacesFromPodAffinityTerm, 2)" -> {"(getIfNewerThan, 2)";}
"(Clear, 0)" -> {"(isUnderlinedHeading, 1)";}
"(wrapperKeyBSONType, 1)" -> {"(dispatch, 0)";"(runSTM, 2)";"(NewClientConnectionWithPings, 2)";}
"(yaml_emitter_delete, 1)" -> {}
"(NewAggregator, 4)" -> {}
"(NewProxier, 13)" -> {}
"(AsFloat, 0)" -> {"(mustWrapValueReflect, 3)";}
"(failure, 1)" -> {"(ToJSONStream, 1)";}
"(callTTBFunc, 3)" -> {}
"(getClientConn, 3)" -> {"(getStartDialLocked, 1)";}
"(ReplaceOrInsert, 1)" -> {"(insert, 2)";}
"(DefineTests, 2)" -> {"(testScriptInPod, 4)";}
"(createProxier, 10)" -> {"(iptablesFlush, 1)";}
"(addServiceOnPortInternal, 4)" -> {"(ProxyLoop, 3)";}
"(monitorProbingStatus, 5)" -> {}
"(freepages, 0)" -> {}
"(sendConsumeCPURequests, 4)" -> {}
"(lockedUpdateLoadBalancerHosts, 2)" -> {"(GetLoadBalancer, 3)";}
"(unsafeCommit, 1)" -> {"(, 2)";}
"(Lock, 0)" -> {"(Ctx, 0)";}
"(newHTTP2Server, 2)" -> {}
"(AddWork, 3)" -> {"(getWrappedWorkerFunc, 1)";}
"(addMemberByURL, 3)" -> {"(listenGRPC, 0)";}
"(Launch, 0)" -> {"(NewClientHandler, 3)";}
"(AttachVolume, 2)" -> {"(GenerateAttachVolumeFunc, 2)";}
"(clientAuthenticate, 1)" -> {"(method, 0)";}
"(Start, 0)" -> {"(walkOperation, 1)";}
"(eof, 0)" -> {"(Detach, 2)";}
"(SetScale, 1)" -> {"(rescale, 1)";}
"(NewProxier, 6)" -> {"(createProxier, 6)";}
"(RuleProtocol, 1)" -> {"(RemoveInstance, 4)";}
"(deadlineLoop, 0)" -> {"(recvKeepAlive, 1)";}
"(handleAuth, 2)" -> {"(authCapabilityHandler, 1)";}
"(newAuthenticator, 2)" -> {"(initVerifier, 3)";}
"(AbsHostPath, 1)" -> {"(NewConnection, 2)";"(newConnection, 4)";}
"(createDialer, 0)" -> {"(ObserveDialLatency, 3)";}
"(waitForA, 1)" -> {"(failure, 1)";}
"(GetIntelRdtPath, 1)" -> {"(getIntelRdtRoot, 0)";}
"(DefaultControllerRateLimiter, 0)" -> {"(NewLimiter, 2)";}
"(applySnapshot, 2)" -> {}
"(generateEvent, 5)" -> {}
"(Unmarshal, 2)" -> {"(unmarshal, 3)";}
"(lookup, 0)" -> {"(compileUpdate, 1)";}
"(findJsonNameField, 1)" -> {"(allocate, 0)";}
"(Id, 1)" -> {"(topIndex, 0)";}
"(yaml_emitter_state_machine, 2)" -> {"(Gather, 0)";"(configureTransports, 1)";"(newWatcherGrpcStream, 1)";"(dispatch, 0)";"(StartRequest, 7)";"(resolveNow, 1)";"(NewClientConnectionWithPings, 2)";"(applyServiceConfigAndBalancer, 2)";}
"(IsUnknown, 1)" -> {}
"(doVolumeSetGetResponse, 2)" -> {"(DiscoverNode, 1)";}
"(GetTestImage, 1)" -> {}
"(NewClient, 2)" -> {"(defaultBaseTransport, 2)";}
"(HcsCloseProcess, 2)" -> {"(hcsCloseProcess, 1)";}
"(HasPrefix, 1)" -> {"(GetPrefix, 1)";}
"(createProxier, 6)" -> {"(platformDoesNotSupportError, 1)";}
"(convertNumber, 1)" -> {}
"(Snapshot, 2)" -> {}
"(monitorResizeEvents, 3)" -> {}
"(Errorf, 4)" -> {"(Error, 3)";}
"(Node, 0)" -> {"(newNode, 1)";}
"(CopyAllLogs, 4)" -> {}
"(NewTimeoutListener, 5)" -> {"(wrapTLS, 3)";}
"(AddInstances, 4)" -> {}
"(defaultBaseTransport, 2)" -> {"(appengineUrlfetchHook, 1)";}
"(writeLines, 3)" -> {"(CloseNotify, 0)";"(on100, 0)";"(SendSnapshot, 1)";"(mergeService, 1)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(newSnapshotReaderCloser, 2)";"(DeleteCollection, 4)";"(handleConnection, 2)";"(SyncServices, 1)";"(PortForward, 3)";"(HandlePodCleanups, 0)";"(startFrameWrite, 1)";"(NewClientConnectionWithPings, 2)";"(processPodOnNode, 5)";"(newDelayingQueue, 3)";"(removeAndReAdd, 1)";"(dispatch, 0)";"(AttachVolume, 2)";"(GenerateRegisterPluginFunc, 4)";}
"(format, 3)" -> {"(writeLines, 3)";}
"(WaitForAttachedDevice, 1)" -> {}
"(ProxyURL, 1)" -> {"(configureHTTP2Transport, 1)";}
"(Nanoseconds, 0)" -> {}
"(Equal, 4)" -> {"(formatUnequalValues, 2)";}
"(ToJSON, 1)" -> {"(SetBuffer, 1)";}
"(FindNamedType, 1)" -> {}
"(openDir, 1)" -> {}
"(tunnelHTTPConnect, 3)" -> {"(startHealthCheck, 1)";}
"(paragraph, 1)" -> {"(isUnderlinedHeading, 1)";}
"(Call, 5)" -> {"(RpcError_ErrorCode, 1)";"(GetDetail, 0)";}
"(updateEffectiveWindowSize, 0)" -> {"(lookup, 0)";}
"(Verify, 2)" -> {"(ecHash, 1)";}
"(Debug, 1)" -> {"(Properties, 4)";}
"(RemoveEndpoints, 3)" -> {"(doReconcile, 3)";}
"(newSnapshotSender, 4)" -> {}
"(Apiextensions, 0)" -> {"(New, 3)";}
"(LinesOf, 1)" -> {"(Line, 0)";}
"(NewDiscoveryController, 3)" -> {"(AddEventHandler, 1)";}
"(updateLastTransitionTime, 3)" -> {"(getIfNewerThan, 2)";}
"(timeoutFlush, 1)" -> {}
"(Logf, 3)" -> {"(releaseEntry, 1)";}
"(AddUint64, 2)" -> {"(AppendUint64, 1)";}
"(New, 6)" -> {"(newWatchableStore, 6)";}
"(canRetryError, 1)" -> {"(getStartDialLocked, 1)";}
"(initVerifier, 3)" -> {"(cancel, 2)";}
"(AfterSuiteDidRun, 1)" -> {"(handleSetupSummary, 2)";}
"(ecHash, 1)" -> {"(fromBase64, 1)";"(packKeyWire, 2)";}
"(Quote, 1)" -> {"(GRPCStatus, 0)";}
"(asyncFlush, 1)" -> {"(SearchFloat64s, 2)";}
"(mergeService, 1)" -> {"(addServicePortPortal, 5)";}
"(startDBSizeMonitorPerEndpoint, 2)" -> {}
"(NewVerifier, 3)" -> {}
"(doVolumeSet, 2)" -> {"(newWatcherGrpcStream, 1)";}
"(Reset, 0)" -> {"(Reset, 1)";"(Inf, 1)";"(MapRange, 0)";}
"(HTTPMembers, 0)" -> {"(isMembersEqual, 2)";}
"(get, 1)" -> {"(MapIndex, 1)";"(toMapKey, 1)";}
"(toMapKey, 1)" -> {"(retrieveUnexportedField, 3)";}
"(AddSpecific, 4)" -> {"(withRetries, 2)";}
"(getContainerStats, 1)" -> {"(Statistics, 0)";}
"(publicJWKSFromKeys, 1)" -> {"(Public, 0)";}
"(newFile, 2)" -> {"(loadPackage, 1)";}
"(NewV1PodResourcesServer, 3)" -> {"(serveUDP, 1)";}
"(walkPaths, 1)" -> {"(walkPathItem, 1)";}
"(enable, 0)" -> {}
"(uint64, 0)" -> {"(ReadUvarint, 1)";}
"(sendMergedSnap, 1)" -> {"(SendSnapshot, 1)";}
"(openExistingOrNew, 1)" -> {"(OpenFile, 3)";}
"(BeforeSuiteDidRun, 1)" -> {"(handleSetupSummary, 2)";}
"(acceptLoop, 0)" -> {}
"(NewEncoder, 1)" -> {"(New, 1)";"(commit, 1)";"(Write, 1)";"(newEncoderWithWriter, 1)";"(marshalDoc, 2)";"(setHttpHeader, 0)";"(NewEncoder, 1)";"(Caller, 1)";"(, 0)";"(Lock, 1)";"(Encode, 2)";"(Put, 1)";"(init, 0)";"(Unlock, 1)";"(Reset, 0)";"(Flush, 0)";"(marshal, 1)";}
"(removeAndReAdd, 1)" -> {}
"(Resolve, 1)" -> {"(FindNamedType, 1)";}
"(portForwardFunc, 3)" -> {"(WriteCloserWrapper, 1)";}
"(toRuleProtocol, 1)" -> {"(RuleProtocol, 1)";}
"(Offset, 1)" -> {"(quoteChar, 1)";}
"(newTimeBudget, 1)" -> {}
"(addPeerToProber, 6)" -> {"(monitorProbingStatus, 5)";}
"(KeysPrefix, 2)" -> {}
"(writeFrameFromHandler, 1)" -> {}
"(TailFile, 2)" -> {"(tailFileSync, 0)";}
"(StrategicMerge, 1)" -> {"(reload, 1)";}
"(newWatcherGrpcStream, 1)" -> {"(GetClientConn, 2)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";}
"(NewVersionedSet, 3)" -> {"(sortByLength, 2)";}
"(doReconcile, 3)" -> {"(Create, 2)";}
"(CancelRequest, 1)" -> {}
"(GetMetricWithLabelValues, 1)" -> {}
"(Install, 1)" -> {"(Consumes, 1)";}
"(Run, 4)" -> {"(runHTTPHandler, 3)";}
"(UpdateTransport, 4)" -> {"(updateTransport, 5)";}
"(panicMessage, 1)" -> {"(typeName, 0)";}
"(From, 0)" -> {"(Wrapf, 2)";}
"(newDecoratedWatcher, 2)" -> {}
"(StartCompactor, 3)" -> {}
"(Format, 4)" -> {"(Debug, 1)";}
"(splitPrefix, 1)" -> {"(newWatcherGrpcStream, 1)";}
"(Run, 3)" -> {"(waitForA, 1)";}
"(typeName, 0)" -> {}
"(ServeConn, 2)" -> {"(serve, 0)";}
"(AddNamedImport, 4)" -> {}
"(CreateVolume, 3)" -> {"(propsSlice, 1)";}
"(Connect, 0)" -> {"(connect, 0)";}
"(resetMessage, 1)" -> {"(SetUnknown, 1)";}
"(validate, 2)" -> {"(versionRange, 1)";}
"(newReadOnly, 1)" -> {"(LastIndex, 0)";"(maybeLastIndex, 0)";}
"(lockAndFlushAll, 0)" -> {"(writeDataFromHandler, 3)";"(newIPVSRequest, 1)";}
"(Panicf, 1)" -> {"(Clone, 0)";}
"(SyncServices, 1)" -> {}
"(InspectExec, 1)" -> {"(ContainerExecInspect, 2)";}
"(GetVMSetNames, 2)" -> {"(getAgentPoolScaleSets, 1)";}
"(stop, 0)" -> {"(DeleteProducer, 1)";}
"(ObserveDialLatency, 3)" -> {"(tunnelHTTPConnect, 3)";}
"(PutResources, 2)" -> {}
"(getJSONEncoder, 0)" -> {"(newWatcherGrpcStream, 1)";"(NewWithDelegate, 1)";"(roundTrip, 1)";"(NewDelayingQueueWithCustomClock, 2)";"(generateEvent, 5)";"(KeepAlive, 2)";"(SendSnapshot, 1)";"(applyServiceConfigAndBalancer, 2)";"(HandleStreams, 2)";"(newSnapshotReaderCloser, 2)";"(CloseNotify, 0)";"(handleConnection, 2)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(ServeConn, 2)";"(resolveNow, 1)";"(listenerRoutine, 0)";"(PortForward, 3)";"(recordToSink, 2)";"(makeConnectedServerPipe, 0)";"(on100, 0)";"(startFrameWrite, 1)";"(awaitOpenSlotForRequest, 1)";"(NewCacherFromConfig, 1)";"(newCluster, 2)";"(newCollector, 4)";"(apply, 3)";"(run, 3)";"(startStreamWriter, 6)";"(AddHTTP, 3)";"(DeleteCollection, 4)";"(NewServerTransport, 3)";"(scheduleBodyWrite, 0)";"(StartRequest, 7)";"(GetClientConn, 2)";}
"(ElectionServerToElectionClient, 1)" -> {"(stopNotify, 0)";}
"(VisitPodConfigmapNames, 2)" -> {"(visitContainerConfigmapNames, 2)";}
"(ProxyLoop, 3)" -> {"(getBackendConn, 7)";}
"(cancel, 0)" -> {"(Stop, 0)";}
"(Run, 2)" -> {"(runRaw, 2)";}
"(printNewLine, 0)" -> {"(tailFileSync, 0)";}
"(newClientConn, 2)" -> {}
"(UpdatePod, 1)" -> {"(syncPodFn, 1)";}
"(MessageOf, 1)" -> {"(Init, 1)";}
"(CompareAndSwapPointer, 3)" -> {}
"(peekType, 0)" -> {"(wrapperKeyBSONType, 1)";}
"(stopDBSizeMonitor, 0)" -> {"(UpdateEtcdDbSize, 2)";}
"(newBridge, 1)" -> {"(serveListen, 0)";}
"(addEventFile, 5)" -> {}
"(newLessor, 3)" -> {}
"(keyFn, 0)" -> {"(AddEventHandlerWithResyncPeriod, 2)";}
"(indirect, 1)" -> {"(eval, 1)";}
"(Statistics, 0)" -> {"(Properties, 2)";}
"(Terminate, 0)" -> {"(Terminate, 1)";}
"(Equal, 3)" -> {"(Equal, 4)";}
"(Delete, 1)" -> {"(Zero, 1)";}
"(reload, 1)" -> {"(recordToSink, 2)";"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(runOnce, 2)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(sendUpdates, 0)";"(NewProxier, 9)";"(ListenAndServe, 8)";"(StartInterceptingOutput, 0)";"(Run, 4)";"(GenerateBulkVolumeVerifyFunc, 4)";"(ListenAndServeReadOnly, 3)";}
"(verify, 2)" -> {"(keysFromRemote, 1)";}
"(newClient, 1)" -> {"(autoSync, 0)";}
"(New, 0)" -> {"(NewFileSet, 0)";}
"(WithStack, 1)" -> {}
"(Watch, 3)" -> {"(newWatcherGrpcStream, 1)";"(Connect, 0)";}
"(addServicePortPortal, 5)" -> {"(ProxyLoop, 3)";}
"(announceSetupFailure, 4)" -> {"(printFailure, 4)";}
"(WriteCloserWrapper, 1)" -> {"(ExecInContainer, 10)";}
"(newEncoderWithWriter, 1)" -> {"(yaml_emitter_set_unicode, 2)";}
"(BrowseDiagnosticLog, 3)" -> {"(newProperties, 1)";}
"(getMetricsFromNode, 2)" -> {}
"(iptablesFlush, 1)" -> {}
"(LeaseKeepAlive, 1)" -> {}
"(endBlock, 0)" -> {"(printDelimiter, 0)";}
"(fail, 2)" -> {"(NewEncoder, 1)";}
"(newProperties, 1)" -> {"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(GetClientConn, 2)";}
"(ServerPreferredResources, 1)" -> {"(fetchGroupVersionResources, 2)";}
"(getBackendConn, 7)" -> {}
"(walkResponse, 1)" -> {"(getJSONEncoder, 0)";}
"(awaitOpenSlotForRequest, 1)" -> {}
"(NewDesc, 4)" -> {"(labelPairSorter, 1)";}
"(closeRemoteChannels, 0)" -> {"(createCall, 5)";"(action, 1)";}
"(GC, 0)" -> {}
"(ForResource, 1)" -> {"(Samplecontroller, 0)";"(Flunders, 0)";}
"(To, 1)" -> {"(do, 3)";}
"(Delete, 2)" -> {"(processRaftRequest, 2)";}
"(doAppendInner, 2)" -> {"(isYesD, 0)";}
"(verifyDisksAttachedToNode, 2)" -> {}
"(escape, 1)" -> {"(printNewLine, 0)";}
"(Zero, 0)" -> {"(MessageOf, 1)";}
"(IsSelected, 1)" -> {"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(StartInterceptingOutput, 0)";"(NewProxier, 9)";"(ListenAndServeReadOnly, 3)";"(recordToSink, 2)";"(runOnce, 2)";"(sendUpdates, 0)";"(ListenAndServe, 8)";}
"(runHTTPHandler, 3)" -> {"(getHTTPRespBody, 1)";}
"(Handle, 2)" -> {"(Head, 2)";}
"(ListPodStatsAndUpdateCPUNanoCoreUsage, 0)" -> {"(listPodStats, 1)";}
"(getContainersToDeleteInPod, 3)" -> {"(updateStatusInternal, 3)";}
"(NewVersion, 1)" -> {"(FromGRPC, 1)";}
"(roundTrip, 1)" -> {"(handleReadLoopResponse, 1)";}
"(isLiteral, 1)" -> {"(cpuMask, 1)";}
"(NewStreamWatcher, 2)" -> {}
"(StartInterceptingOutput, 0)" -> {"(TailFile, 2)";}
"(StartRequest, 7)" -> {}
"(AnnounceSpecRunCompletion, 2)" -> {}
"(printFailure, 4)" -> {"(waitForChanges, 0)";}
"(processRaftRequest, 2)" -> {"(newWatcherGrpcStream, 1)";"(Connect, 0)";"(markDiskAsDeletable, 2)";"(parseProposeCtxErr, 2)";}
"(writeDataFromHandler, 3)" -> {"(writeFrameFromHandler, 1)";"(SortStrings, 1)";}
"(failureDetails, 1)" -> {"(endBlock, 0)";"(printSpecFailure, 4)";}
"(ioctlSetInt, 3)" -> {"(registerEvent, 3)";}
"(WithDeadline, 2)" -> {"(cancel, 2)";}
"(emitScalar, 8)" -> {"(callK, 1)";"(handleInterrupt, 0)";"(ConfigureTransports, 1)";}
"(SetError, 0)" -> {"(addEvent, 3)";}
"(SetBuffer, 1)" -> {"(constructBlock, 1)";}
"(ListenForSignal, 1)" -> {}
"(VolumesInUse, 2)" -> {"(volumesInUseFunc, 0)";}
"(unmarshal, 3)" -> {"(newParser, 1)";"(unmarshalPointer, 4)";}
"(Warn, 5)" -> {"(compact, 3)";}
"(isFeatureInRange, 2)" -> {}
"(errorS, 5)" -> {"(Insert, 1)";"(Remove, 1)";"(Do, 1)";"(Add, 1)";"(pathsToRemove, 2)";"(Add, 2)";"(Wait, 0)";"(Update, 3)";"(Reset, 0)";"(Flush, 0)";"(Transform, 3)";"(Send, 1)";"(openExistingOrNew, 1)";"(rotate, 0)";"(Header, 0)";"(append, 1)";"(Copy, 2)";"(New, 3)";"(Parse, 1)";"(Sync, 0)";"(Encode, 2)";"(Call, 5)";"(New, 0)";"(Decode, 1)";"(V, 1)";}
"(Set, 2)" -> {"(update, 4)";}
"(getStartDialLocked, 1)" -> {}
"(NormalizedName, 1)" -> {"(verify, 1)";}
"(fillClusterForMembers, 0)" -> {"(listenGRPC, 0)";}
"(msync, 1)" -> {"(newProxyServer, 3)";"(createCall, 5)";"(dialStreamLocal, 1)";"(monitorSize, 2)";"(Launch, 1)";"(createAndInitKubelet, 30)";"(SpecSuiteWillBegin, 2)";"(SpecSuiteDidEnd, 1)";"(hive, 1)";"(newConn, 2)";"(startKubelet, 6)";"(NewAggregator, 4)";"(RunOnce, 1)";}
"(NewEstablishingController, 2)" -> {"(DefaultControllerRateLimiter, 0)";}
"(fetchContainerStats, 1)" -> {"(Statistics, 0)";}
"(hcsCloseComputeSystem, 1)" -> {}
"(run, 3)" -> {"(StartGarbageCollection, 0)";"(sendUpdates, 0)";"(StartInterceptingOutput, 0)";"(CreateHealthCheck, 1)";"(NewTimeoutListener, 5)";"(ListenAndServeReadOnly, 3)";"(monitorResizeEvents, 3)";"(startFrameWrite, 1)";"(CloseNotify, 0)";"(AttachVolume, 2)";"(PortForward, 3)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(VerifySignature, 2)";"(Eventf, 7)";"(newSnapshotReaderCloser, 2)";"(runCmdContext, 2)";"(GenerateRegisterPluginFunc, 4)";"(createVolume, 2)";"(processPodOnNode, 5)";"(runOnce, 2)";"(Launch, 0)";"(on100, 0)";"(SendSnapshot, 1)";"(NewProxier, 9)";"(handleConnection, 2)";"(serve, 0)";"(Dispatch, 4)";"(HandlePodCleanups, 0)";"(verify, 1)";"(ListenAndServe, 8)";"(removeAndReAdd, 1)";"(SyncServices, 1)";"(lockedUpdateLoadBalancerHosts, 2)";"(generateEvent, 5)";"(NewMainKubelet, 30)";"(mergeService, 1)";"(createDialer, 0)";}
"(tearDown, 1)" -> {}
"(unlock, 0)" -> {"(ref, 0)";}
"(InitService, 1)" -> {}
"(handleCall, 1)" -> {"(decodeArguments, 3)";}
"(ListenAndServe, 0)" -> {"(serveUDP, 1)";}
"(newReflectCache, 0)" -> {"(Store, 1)";"(AsFloat, 0)";"(NewUnstructuredObjectTyper, 0)";}
"(NewSessionWithOptions, 1)" -> {"(newSession, 3)";}
"(resolveNow, 1)" -> {}
"(init, 0)" -> {"(AddExtensionType, 2)";}
"(iterate, 6)" -> {}
"(Update, 3)" -> {"(createMetadatas, 1)";}
"(processSignalResult, 2)" -> {}
"(cpuMask, 1)" -> {"(Make, 1)";}
"(truncatingFormat, 1)" -> {"(callTTBFunc, 3)";"(callTRFunc, 3)";}
"(MapRange, 0)" -> {"(closeRemoteChannels, 0)";}
"(NewWithDelegate, 1)" -> {}
"(New, 2)" -> {"(format, 2)";}
"(GracefulStop, 0)" -> {}
"(RoundTripOpt, 2)" -> {"(shouldRetryRequest, 3)";}
"(watchErrorStream, 2)" -> {}
"(mountServiceAccountToken, 2)" -> {"(createVolume, 2)";}
"(GetFullQualifiedPluginNameForVolume, 2)" -> {"(setNodeVolume, 4)";}
"(mustNewMember, 1)" -> {"(listenGRPC, 0)";}
"(newLoaderAtConfirmedDir, 4)" -> {"(HasPrefix, 1)";}
"(Create, 2)" -> {"(CreateUser, 3)";}
"(inBounds, 3)" -> {"(Run, 4)";"(runOnce, 2)";"(ListenAndServeReadOnly, 3)";"(handleConnection, 2)";"(cleanupOrphanedPodCgroups, 3)";"(GenerateRegisterPluginFunc, 4)";"(ListenAndServe, 8)";"(initForOS, 2)";"(addServiceOnPortInternal, 4)";"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(monitorResizeEvents, 3)";"(NewClientConnectionWithPings, 2)";"(createVolume, 2)";"(addServicePortPortal, 5)";"(sendUpdates, 0)";"(dispatch, 0)";"(Eventf, 7)";"(processPodOnNode, 5)";"(VerifySignature, 2)";"(NewClusterV3, 2)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(inBounds, 3)";"(RegisterPlugin, 3)";"(startDBSizeMonitorPerEndpoint, 2)";"(HandlePodCleanups, 0)";"(NewProxier, 9)";"(recordToSink, 2)";"(WaitForAttachedDevice, 1)";"(AddWork, 3)";"(startCompactorOnce, 2)";"(mergeService, 1)";"(UpdatePod, 1)";"(Launch, 0)";"(StartInterceptingOutput, 0)";"(Dispatch, 4)";"(verify, 1)";"(removeAndReAdd, 1)";"(SyncServices, 1)";"(Chan, 0)";"(AttachVolume, 2)";}
"(newSnapshotReaderCloser, 2)" -> {}
"(Execute, 2)" -> {"(Receive, 0)";}
"(rotate, 0)" -> {"(mill, 0)";}
"(After, 1)" -> {"(unlock, 0)";}
"(equalMessage, 2)" -> {"(equalUnknown, 2)";}
"(Header, 0)" -> {"(FromContext, 1)";"(finish, 1)";}
"(printf, 5)" -> {"(output, 6)";}
"(ServiceTicket, 3)" -> {"(SetThumbprint, 2)";}
"(labelPairSorter, 1)" -> {"(doAppendInner, 2)";}
"(newStore, 1)" -> {"(NewUnsafeSet, 1)";}
"(MigrateIfNeeded, 1)" -> {"(rollbackEtcd3MinorVersion, 2)";}
"(OperationCompleteHook, 2)" -> {"(setNodeVolume, 4)";}
"(Start, 1)" -> {"(walkPaths, 1)";}
"(Reset, 1)" -> {"(ProtoReflect, 0)";"(resetMessage, 1)";}
"(CreateUser, 3)" -> {"(roundTrip, 1)";"(startCompactorOnce, 2)";"(RoundTripOpt, 2)";"(UpdatePod, 1)";"(startDBSizeMonitorPerEndpoint, 2)";}
"(UpdateEtcdDbSize, 2)" -> {}
"(checkForUpdates, 1)" -> {"(getIfNewerThan, 2)";}
"(HcsCloseComputeSystem, 2)" -> {"(hcsCloseComputeSystem, 1)";}
"(Lock, 1)" -> {"(Ctx, 0)";"(waitDeletes, 4)";}
"(NewError, 3)" -> {"(validate, 2)";}
"(NewCRDFinalizer, 3)" -> {"(AddEventHandler, 1)";}
"(NewDiscoveryRESTMapper, 1)" -> {"(AddSpecific, 4)";}
"(Close, 0)" -> {"(Send, 1)";}
"(runRaw, 2)" -> {"(runWithFriendlyError, 3)";}
"(valueEncoder, 1)" -> {"(typeEncoder, 1)";}
"(addConnIfNeeded, 3)" -> {}
"(ExecInContainer, 10)" -> {"(InspectExec, 1)";}
"(Kill, 1)" -> {"(init, 0)";}
"(WaitForSingleObject, 2)" -> {}
"(Receive, 0)" -> {"(ParseNetlinkMessage, 1)";}
"(NewUnstructuredObjectTyper, 0)" -> {"(waitCancelSubstreams, 1)";"(getMetricsFromNode, 2)";"(newSession, 2)";"(DefineTests, 2)";"(SetUpAt, 2)";"(MigrateIfNeeded, 1)";"(newFilePipeline, 3)";"(newDecoratedWatcher, 2)";"(newConn, 2)";"(createMergedSnapshotMessage, 4)";"(PortForward, 4)";"(UpdateTransport, 4)";"(AddPod, 1)";"(BuildAuthn, 2)";"(receive, 0)";"(createCall, 5)";"(Evaluate, 1)";"(SpecSuiteDidEnd, 1)";"(StartTestServerOrDie, 4)";"(sendMergedSnap, 1)";"(startKubelet, 6)";"(CopyAllLogs, 4)";"(WatchPods, 4)";"(handleResponse, 2)";"(StartGatheringData, 0)";"(Dgemm, 13)";"(mountServiceAccountToken, 2)";"(passesVerifyFilter, 2)";"(copyStdout, 1)";"(RediscoverNode, 1)";"(copyStderr, 1)";"(watchErrorStream, 2)";"(Sink, 0)";"(handleResizes, 0)";"(createAndInitKubelet, 30)";"(freepages, 0)";"(NewStreamWatcher, 2)";"(Build, 3)";"(scheduleFrameWrite, 0)";"(DialWithConn, 4)";"(NewProxier, 13)";"(NewAggregator, 4)";"(initializeCSINode, 1)";"(waitForAvailableEndpoint, 1)";"(NewClientConn, 3)";"(newProxyServer, 3)";"(ListenForSignal, 1)";"(applyEntries, 2)";"(copyStdin, 0)";"(inWorker, 0)";"(monitorSize, 2)";"(Launch, 1)";"(Dispatch, 3)";"(RunOnce, 1)";"(SpecSuiteWillBegin, 2)";"(dialStreamLocal, 1)";"(createPD, 1)";"(recvLoop, 0)";"(podsRunning, 2)";"(newETCD3Storage, 2)";"(Verify, 2)";"(applySnapshot, 2)";"(sendLoop, 0)";}
"(eval, 1)" -> {"(getRequestedContainers, 2)";}
"(HandlePodCleanups, 0)" -> {"(GC, 0)";}
"(NewListener, 2)" -> {"(NewTimeoutListener, 5)";}
"(SyncPod, 4)" -> {"(containerStartSpec, 1)";}
"(log, 4)" -> {"(sweetenFields, 1)";}
"(fetch, 1)" -> {"(GetResponseRange, 0)";}
"(yaml_emitter_set_unicode, 2)" -> {"(TrimRightFunc, 2)";}
"(Get, 0)" -> {"(string, 1)";"(GetLatestPage, 0)";}
"(Errorf, 5)" -> {"(Warn, 5)";}
"(Sub, 1)" -> {"(AddUint64, 2)";}
"(newParser, 1)" -> {"(yaml_parser_set_input_string, 2)";}
"(recvKeepAlive, 1)" -> {"(sendKeepAliveLoop, 1)";}
"(newProxyServer, 3)" -> {"(NewProxier, 6)";}
"(NewFileSet, 0)" -> {"(Exec, 2)";"(OptionManager, 0)";}
"(mill, 0)" -> {}
"(startFrameWrite, 1)" -> {}
"(HandlePodReconcile, 1)" -> {"(deleteContainersInPod, 3)";}
"(markDiskAsDeletable, 2)" -> {"(Upload, 4)";}
"(OpenFile, 3)" -> {}
"(GetResponseRange, 0)" -> {"(checkTxnAuth, 3)";}
"(CanSet, 0)" -> {"(Evaluate, 1)";"(Dispatch, 3)";"(roundTrip, 1)";"(RoundTripOpt, 2)";"(newConn, 2)";"(dialStreamLocal, 1)";}
"(unregisterCallback, 1)" -> {"(HcsUnregisterComputeSystemCallback, 2)";}
"(kill, 0)" -> {"(Errorf, 5)";}
"(destroy, 0)" -> {"(yaml_emitter_delete, 1)";}
"(serve, 0)" -> {"(shutDownIn, 1)";}
"(ListenAndServe, 8)" -> {"(ListenAndServeKubeletServer, 10)";}
"(appengineUrlfetchHook, 1)" -> {"(FindDefaultCredentials, 2)";}
"(FindDefaultCredentials, 2)" -> {"(newSnapshotReaderCloser, 2)";"(CloseNotify, 0)";"(on100, 0)";"(startFrameWrite, 1)";"(PortForward, 3)";"(SendSnapshot, 1)";}
"(Detach, 2)" -> {"(doVolumeSet, 2)";}
"(addLivezChecks, 2)" -> {"(processPodOnNode, 5)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(handleConnection, 2)";"(SyncServices, 1)";"(HandlePodCleanups, 0)";"(mergeService, 1)";"(removeAndReAdd, 1)";"(AttachVolume, 2)";"(GenerateRegisterPluginFunc, 4)";}
"(Init, 1)" -> {"(CompareAndSwapPointer, 3)";}
"(waitForAvailableEndpoint, 1)" -> {}
"(receive, 0)" -> {}
"(handleInterrupt, 0)" -> {}
"(updateSecurityGroup, 4)" -> {"(createNodeSecurityGroup, 5)";}
"(popError, 1)" -> {"(WithLabelValues, 1)";}
"(HcsUnregisterComputeSystemCallback, 2)" -> {"(hcsUnregisterComputeSystemCallback, 1)";}
"(satisfiedExpectations, 0)" -> {"(poller, 2)";}
"(DeliverSignal, 3)" -> {"(deliver, 1)";}
"(RemoveContainer, 1)" -> {"(popError, 1)";}
"(cadvisorInfoToAcceleratorStats, 1)" -> {"(getRequestedContainers, 2)";"(OpenContainer, 1)";"(fetchContainerStats, 1)";"(getContainerStats, 1)";}
"(requestLogger, 2)" -> {"(handleAuth, 2)";}
"(wordMask, 1)" -> {"(effectivelyEquals, 1)";}
"(FileLine, 1)" -> {"(DialWithConn, 4)";}
"(maybeLastIndex, 0)" -> {"(RoundTripOpt, 2)";"(dispatch, 0)";"(runSTM, 2)";"(NewClientConnectionWithPings, 2)";"(roundTrip, 1)";}
"(DrainedAt, 0)" -> {"(RemoveImage, 1)";}
"(sortByLength, 2)" -> {}
"(ServerGroupsAndResources, 1)" -> {"(fetchGroupVersionResources, 2)";}
"(initForOS, 2)" -> {"(InitService, 1)";}
"(ChangeEvents, 2)" -> {}
"(do, 3)" -> {"(ErrAllFailed, 1)";}
"(SearchFloat64s, 2)" -> {"(Run, 4)";"(newWatcherGrpcStream, 1)";"(BulkVerifyVolumes, 1)";"(initForOS, 2)";}
"(QueryServiceConfig2, 5)" -> {"(clientConfig, 1)";}
"(newIPVSRequest, 1)" -> {"(writeLines, 3)";}
"(StartTestServer, 4)" -> {}
"(Gather, 0)" -> {}
"(StringAffineComparable, 1)" -> {"(doVolumeSetGetResponse, 2)";}
"(V1, 0)" -> {"(New, 3)";}
"(GetFileType, 1)" -> {"(getFileType, 1)";}
"(addCertRotation, 6)" -> {}
"(NewUnsecuredEtcd3TestClientServer, 1)" -> {"(PathPrefix, 0)";}
"(RegisterPlugin, 3)" -> {}
"(applyServiceConfigAndBalancer, 2)" -> {"(newCCBalancerWrapper, 3)";}
"(output, 1)" -> {"(objectType, 1)";}
"(sendConsumeCustomMetric, 5)" -> {}
"(syncPodFn, 1)" -> {"(wrapUp, 2)";}
"(ListenAndServePodResources, 4)" -> {"(NewV1PodResourcesServer, 3)";}
"(Now, 0)" -> {"(TimeFromUnixNano, 1)";}
"(applyEntries, 2)" -> {}
"(ListenAndServeReadOnly, 3)" -> {"(ListenAndServeKubeletReadOnlyServer, 5)";}
"(transformStartContainerError, 1)" -> {}
"(objectType, 1)" -> {"(dialStreamLocal, 1)";"(newConn, 2)";}
"(createPD, 1)" -> {"(CreatePD, 1)";}
"(CreatePD, 1)" -> {"(CreateDisk, 5)";}
"(selectLoadBalancer, 4)" -> {"(GetVMSetNames, 2)";}
"(GetPrefix, 1)" -> {"(Gather, 0)";"(Evaluate, 1)";"(StartTestServerOrDie, 4)";"(Dgemm, 13)";"(markDiskAsDeletable, 2)";"(mountServiceAccountToken, 2)";"(processPodOnNode, 5)";"(applySnapshot, 2)";"(handleResponse, 2)";"(watchErrorStream, 2)";"(createCall, 5)";"(newETCD3Storage, 2)";"(ConfigureTransports, 1)";"(Verify, 2)";"(newProxyServer, 3)";"(NewClientConn, 3)";"(NewListener, 2)";"(recvLoop, 0)";"(newConn, 2)";"(receive, 0)";"(addPeerToProber, 6)";"(ServeConn, 2)";"(createAndInitKubelet, 30)";"(newHTTP2Transport, 2)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(ListenForSignal, 1)";"(WatchPods, 4)";"(scheduleFrameWrite, 0)";"(Dispatch, 3)";"(ForResource, 1)";"(addMember, 1)";"(initializeCSINode, 1)";"(SetUpAt, 2)";"(run, 3)";"(NewAggregator, 4)";"(startPeer, 4)";"(createDialer, 0)";"(createPD, 1)";"(SpecSuiteDidEnd, 1)";"(mergeService, 1)";"(NewCacherFromConfig, 1)";"(resolveNow, 1)";"(HandlePodUpdates, 1)";"(CreateHealthCheck, 1)";"(startKubelet, 6)";"(removeAndReAdd, 1)";"(podsRunning, 2)";"(SpecSuiteWillBegin, 2)";"(RediscoverNode, 1)";"(RunOnce, 1)";"(GenerateRegisterPluginFunc, 4)";"(NewWithDelegate, 1)";"(handleConnection, 2)";"(monitorSize, 2)";"(Launch, 1)";"(handleResizes, 0)";"(getMetricsFromNode, 2)";"(callK, 1)";"(NewProxier, 13)";"(newDecoratedWatcher, 2)";"(NewClientConnectionWithPings, 2)";"(PortForward, 4)";"(copyStderr, 1)";"(AttachVolume, 2)";"(generateEvent, 5)";"(DeleteCollection, 4)";"(AddPod, 1)";"(Sink, 0)";"(newSession, 2)";"(BuildAuthn, 2)";"(sendLoop, 0)";"(Build, 3)";"(HandlePodReconcile, 1)";"(applyEntries, 2)";"(Run, 4)";"(newFilePipeline, 3)";"(freepages, 0)";"(copyStdin, 0)";"(copyStdout, 1)";"(passesVerifyFilter, 2)";"(applyServiceConfigAndBalancer, 2)";"(UpdateTransport, 4)";"(recordToSink, 2)";"(waitCancelSubstreams, 1)";"(dialStreamLocal, 1)";"(dispatch, 0)";"(NewDelayingQueueWithCustomClock, 2)";"(HandlePodCleanups, 0)";"(createMergedSnapshotMessage, 4)";"(MigrateIfNeeded, 1)";"(SyncServices, 1)";"(NewStreamWatcher, 2)";"(StartGatheringData, 0)";"(waitForAvailableEndpoint, 1)";"(CopyAllLogs, 4)";"(DefineTests, 2)";"(lockedUpdateLoadBalancerHosts, 2)";"(newWatcherGrpcStream, 1)";"(inWorker, 0)";"(listenerRoutine, 0)";"(sendMergedSnap, 1)";}
"(updateClientConnState, 1)" -> {"(applyServiceConfigAndBalancer, 2)";"(switchBalancer, 1)";"(newCCBalancerWrapper, 3)";}
"(checkTxnReqsPermission, 3)" -> {"(Connect, 0)";}
"(propsSlice, 1)" -> {}
"(Fatalf, 2)" -> {"(get, 0)";"(log, 4)";"(printf, 5)";"(FatalDepth, 2)";"(Logf, 3)";"(header, 2)";"(cancel, 0)";"(fail, 2)";"(Exit, 1)";}
"(NewTimer, 1)" -> {"(Now, 0)";}
"(getIntelRdtRoot, 0)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(updateTransport, 5)" -> {"(addCertRotation, 6)";}
"(close, 0)" -> {"(eof, 0)";}
"(newSession, 2)" -> {}
"(compileUpdate, 1)" -> {"(ParseServiceConfig, 1)";"(newWatcherGrpcStream, 1)";}
"(podsRunning, 2)" -> {}
"(sendUpdates, 0)" -> {}
"(setHttpHeader, 0)" -> {"(string, 1)";}
"(addMember, 1)" -> {"(HTTPMembers, 0)";"(addMemberByURL, 3)";}
"(Do, 3)" -> {"(debugResponse, 1)";}
"(newCallSet, 0)" -> {"(NewFileSet, 0)";}
"(waitDeletes, 4)" -> {"(waitDelete, 4)";}
"(GetUnknown, 0)" -> {"(createCall, 5)";"(freepages, 0)";}
"(inWorker, 0)" -> {"(handleCall, 1)";}
"(latestContainerStats, 1)" -> {"(UserDefinedMetricType, 1)";}
"(startCompactorOnce, 2)" -> {"(StartCompactor, 3)";}
"(Verifier, 1)" -> {"(NewVerifier, 3)";}
"(Marshal, 1)" -> {"(destroy, 0)";}
"(NewLimiter, 2)" -> {"(NewDelayingQueueWithCustomClock, 2)";}
"(toStringSlice, 1)" -> {"(QueryServiceConfig2, 5)";}
"(callK, 1)" -> {"(ClockWait, 3)";}
"(packKeyWire, 2)" -> {"(keysFromRemote, 1)";}
"(node, 2)" -> {"(emitScalar, 8)";}
"(format, 4)" -> {"(GRPCStatus, 0)";}
"(VerifySignature, 2)" -> {"(verify, 2)";}
"(SendMsg, 1)" -> {"(outPayload, 5)";}
"(GetFromPath, 1)" -> {"(NsHandle, 1)";}
"(Warningln, 2)" -> {"(newHTTP2Server, 2)";}
"(newHTTP2Client, 7)" -> {"(connectionErrorf, 5)";}
"(print, 4)" -> {"(LinesOf, 1)";"(Strings, 1)";"(String, 0)";"(runHandler, 1)";"(From, 1)";"(NodesOf, 1)";"(From, 0)";"(Infof, 2)";"(String, 2)";"(New, 1)";"(flushAll, 0)";"(add, 2)";"(Version, 0)";}
"(Lstat, 1)" -> {"(NewUnstructuredObjectTyper, 0)";}
"(byPrefixPriority, 1)" -> {}
"(newIntSet, 0)" -> {"(Max, 0)";}
"(sendSnap, 1)" -> {}
"(versionRange, 1)" -> {}
"(removeChild, 2)" -> {"(configureTransports, 1)";}
"(fromBase64, 1)" -> {"(keysFromRemote, 1)";}
"(hcsUnregisterComputeSystemCallback, 1)" -> {}
"(newETCD3Storage, 2)" -> {"(stopDBSizeMonitor, 0)";}
"(typeEncoder, 1)" -> {"(enable, 0)";}
"(newNode, 1)" -> {"(splitPrefix, 1)";}
"(finish, 1)" -> {"(SetError, 0)";}
"(ParseServiceConfig, 1)" -> {"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(sendConsumeCPURequests, 4)";}
"(NewServerTransport, 3)" -> {"(newHTTP2Server, 2)";}
"(setup, 2)" -> {"(ioctlSetInt, 3)";}
"(Run, 1)" -> {"(storeScratch, 4)";}
"(reset, 0)" -> {"(updateEffectiveWindowSize, 0)";}
"(deliver, 1)" -> {}
"(volumesInUseFunc, 0)" -> {}
"(getHTTPRespBody, 1)" -> {}
"(run, 0)" -> {"(processData, 0)";}
"(OptionManager, 0)" -> {"(newDelayingQueue, 3)";"(Example, 0)";"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(addListener, 1)" -> {"(newDelayingQueue, 3)";}
"(String, 2)" -> {"(grow, 2)";}
"(Chan, 0)" -> {}
"(walkOperation, 1)" -> {"(walkResponse, 1)";}
"(header, 2)" -> {"(formatHeader, 3)";}
"(NewCtxClient, 1)" -> {"(WithCancel, 1)";}
"(runCmdContext, 2)" -> {}
"(newCollector, 4)" -> {"(mapEventsToCustomEvents, 1)";}
"(getIfNewerThan, 2)" -> {}
"(InjectHandlers, 1)" -> {"(PushFrontNamed, 1)";}
"(setState, 1)" -> {}
"(DiscoverNode, 1)" -> {}
"(SortableMemberSliceByPeerURLs, 1)" -> {"(handleAuth, 2)";}
"(Get, 3)" -> {"(Do, 3)";}
"(retrieveUnexportedField, 3)" -> {"(UnsafeAddr, 0)";}
"(fdatasync, 1)" -> {"(msync, 1)";}
"(NewDelayingQueueWithCustomClock, 2)" -> {"(newDelayingQueue, 3)";}
"(SpecSuiteWillBegin, 2)" -> {"(escape, 1)";}
"(setNodeVolume, 4)" -> {"(verifyDisksAttachedToNode, 2)";}
"(printSpecFailure, 4)" -> {"(printFailure, 4)";}
"(keyToFilename, 1)" -> {"(Sum, 1)";}
"(Unknown, 1)" -> {"(IsUnknown, 1)";}
"(equalUnknown, 2)" -> {"(AttachVolume, 2)";"(handleConnection, 2)";"(NewDelayingQueueWithCustomClock, 2)";"(newWatcherGrpcStream, 1)";"(mergeService, 1)";"(removeAndReAdd, 1)";"(configureTransports, 1)";"(HandlePodCleanups, 0)";"(GenerateRegisterPluginFunc, 4)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(recordToSink, 2)";"(SyncServices, 1)";"(Run, 4)";"(processPodOnNode, 5)";}
"(configureTransports, 1)" -> {"(addConnIfNeeded, 3)";}
"(createMergedSnapshotMessage, 4)" -> {"(NewMessage, 3)";}
"(Public, 0)" -> {}
"(recordToSink, 2)" -> {}
"(scheduleFrameWrite, 0)" -> {"(isControl, 0)";}
"(handleSetupSummary, 2)" -> {"(failureDetails, 1)";"(announceSetupFailure, 4)";}
"(verify, 1)" -> {"(dedup, 1)";}
"(Retrieve, 0)" -> {"(profile, 0)";}
"(NewWithTLSConfig, 2)" -> {"(ProxyURL, 1)";}
"(mustWrapValueReflect, 3)" -> {"(runSTM, 2)";"(runOnce, 2)";"(newMux, 1)";"(StartTestServer, 4)";"(sendConsumeCPURequests, 4)";"(HandleStreams, 2)";"(newSnapshotReaderCloser, 2)";"(on100, 0)";"(rollbackEtcd3MinorVersion, 2)";"(NewClientConnectionWithPings, 2)";"(listenerRoutine, 0)";"(RoundTripOpt, 2)";"(newWatcherGrpcStream, 1)";"(monitorResizeEvents, 3)";"(SendSnapshot, 1)";"(NewProxier, 9)";"(Chan, 0)";"(CloseNotify, 0)";"(HandlePodCleanups, 0)";"(startFrameWrite, 1)";"(StartNode, 2)";"(DiscoverNode, 1)";"(NewDelayingQueueWithCustomClock, 2)";"(ListenAndServeReadOnly, 3)";"(Eventf, 7)";"(addPeerToProber, 6)";"(newHTTP2Transport, 2)";"(enable, 0)";"(NewMainKubelet, 30)";"(sendUpdates, 0)";"(sendConsumeMemRequests, 4)";"(processPodOnNode, 5)";"(VerifySignature, 2)";"(removeAndReAdd, 1)";"(newClientConn, 2)";"(DeleteCollection, 4)";"(StartGarbageCollection, 0)";"(startPeer, 4)";"(NewSession, 2)";"(sendConsumeCustomMetric, 5)";"(dgemmParallel, 12)";"(CreatePD, 1)";"(updateTransport, 5)";"(CreateVolume, 3)";"(handleConnection, 2)";"(startCompactorOnce, 2)";"(PortForward, 3)";"(UpdatePod, 1)";"(startDBSizeMonitorPerEndpoint, 2)";"(RestartNode, 1)";"(newLessor, 3)";"(mergeService, 1)";"(recordToSink, 2)";"(Run, 4)";"(apply, 3)";"(SyncServices, 1)";"(Launch, 0)";"(StartInterceptingOutput, 0)";"(Dispatch, 4)";"(keepAliveLoop, 2)";"(clientHandshake, 2)";"(verify, 1)";"(GetCollector, 1)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(dispatch, 0)";"(ListenAndServe, 8)";"(AttachVolume, 2)";"(GenerateRegisterPluginFunc, 4)";"(handleSignal, 1)";"(roundTrip, 1)";"(createVolume, 2)";}
"(apply, 2)" -> {"(Panicf, 1)";}
"(NewReader, 4)" -> {"(newBpsLoop, 1)";}
"(TerminatePod, 1)" -> {"(updateStatusInternal, 3)";}
"(healthCheckFunc, 4)" -> {"(NewClientTransport, 7)";}
"(GetLoadBalancer, 3)" -> {"(getFloatingIPByPortID, 2)";}
"(add, 2)" -> {"(Add, 2)";}
"(WithCancel, 1)" -> {"(cancel, 2)";}
"(logName, 2)" -> {"(sendConsumeCPURequests, 4)";"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(runSTM, 2)";}
"(NewClusterAuthenticationTrustController, 2)" -> {"(keyFn, 0)";}
"(clientConfig, 1)" -> {"(connect, 0)";"(newETCD3HealthCheck, 1)";}
"(Store, 1)" -> {"(RetrieveOne, 4)";}
"(Bytes, 0)" -> {"(panicMessage, 1)";}
"(Make, 1)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(Consumes, 1)" -> {"(refreshMuxLocked, 0)";}
"(newMux, 1)" -> {"(newCond, 0)";}
"(NewMessage, 3)" -> {}
"(mapEventsToCustomEvents, 1)" -> {"(setup, 2)";}
"(Kill, 0)" -> {"(convertProcessError, 2)";}
"(openIDKeysetJSON, 1)" -> {"(publicJWKSFromKeys, 1)";}
"(dialStreamLocal, 1)" -> {}
"(createCall, 5)" -> {}
"(Samplecontroller, 0)" -> {}
"(newDecoder, 0)" -> {"(convertSliceNumbers, 2)";}
"(WithTTL, 1)" -> {"(mkSTM, 2)";}
"(PushFrontNamed, 1)" -> {}
"(LoadSources, 3)" -> {"(newFile, 2)";}
"(action, 1)" -> {}
"(method, 0)" -> {}
"(stopNotify, 0)" -> {}
"(Get, 2)" -> {"(reset, 2)";}
"(TrimRightFunc, 2)" -> {"(handleInterrupt, 0)";"(ConfigureTransports, 1)";"(callK, 1)";}
"(ContainerExecInspect, 2)" -> {}
"(newConn, 2)" -> {}
"(SendSnapshot, 1)" -> {"(sendSnap, 1)";}
"(Wait, 0)" -> {"(init, 0)";"(WaitForSingleObject, 2)";}
"(Parse, 1)" -> {"(ReduceFunc, 1)";}
"(Build, 2)" -> {}
"(enableCSM, 3)" -> {"(InjectHandlers, 1)";}
"(NodesOf, 1)" -> {"(Node, 0)";}
"(grow, 2)" -> {"(DecodeRuneInString, 1)";}
"(withRetries, 2)" -> {"(fetchGroupVersionResources, 2)";}
"(GetWorkerUnsafe, 1)" -> {"(getWrappedWorkerFunc, 1)";}
"(FromGRPC, 1)" -> {"(WithStack, 1)";}
"(runWithFriendlyError, 3)" -> {"(run, 3)";}
"(RawSyscallNoError, 4)" -> {"(RemoteCredProvider, 2)";}
"(tailFileSync, 0)" -> {"(waitForChanges, 0)";}
"(Initialize, 2)" -> {"(rebuild, 2)";}
"(InstallAPIGroups, 1)" -> {}
"(NewAddress, 1)" -> {"(updateResolverState, 2)";}
"(profile, 0)" -> {}
"(newETCD3HealthCheck, 1)" -> {}
"(Difference, 1)" -> {"(NewVersionedSet, 3)";}
"(WithLazyPrefix, 1)" -> {"(freepages, 0)";"(markDiskAsDeletable, 2)";}
"(connectionErrorf, 5)" -> {}
"(walkPathItem, 1)" -> {"(walkParameters, 1)";}
"(UnsafeAddr, 0)" -> {"(configureTransports, 1)";"(Upload, 4)";}
"(Zero, 1)" -> {"(CanSet, 0)";}
"(SetTypeForKey, 2)" -> {}
"(NewConditionController, 2)" -> {"(AddEventHandler, 1)";}
"(Stop, 0)" -> {"(New, 1)";"(Kill, 0)";"(Error, 0)";"(RemoveEndpoints, 3)";"(Do, 1)";"(Done, 0)";"(Close, 0)";"(kill, 0)";"(, 0)";"(stop, 0)";"(Kill, 1)";"(Signal, 1)";"(RemoveContainer, 1)";"(Wait, 0)";"(String, 0)";"(Unlock, 0)";"(NewTimer, 1)";"(Lock, 0)";"(Finish, 0)";"(After, 1)";"(Remove, 1)";"(Destroy, 0)";"(V, 1)";}
"(Iterate, 1)" -> {"(IterateUsing, 2)";}
"(runSTM, 2)" -> {}
"(Read, 1)" -> {"(Sum, 1)";}
"(rescale, 1)" -> {"(RediscoverNode, 1)";"(roundTrip, 1)";"(RoundTripOpt, 2)";"(Quo, 2)";"(newConn, 2)";"(dialStreamLocal, 1)";"(freepages, 0)";"(newWatcherGrpcStream, 1)";}
"(Max, 0)" -> {"(max, 1)";}
"(loadPackage, 1)" -> {"(NewChecker, 4)";}
"(remove, 1)" -> {"(wordMask, 1)";}
"(BuildAuthn, 2)" -> {}
"(AddEventHandlerWithResyncPeriod, 2)" -> {"(addListener, 1)";}
"(newConnection, 4)" -> {"(sendPings, 1)";}
"(getRequest, 1)" -> {"(newWatcherGrpcStream, 1)";"(Connect, 0)";}
"(KeepAlive, 2)" -> {"(deadlineLoop, 0)";}
"(WatchPods, 4)" -> {}
"(marshal, 1)" -> {"(reflectValue, 1)";}
"(NewClientConnectionWithPings, 2)" -> {"(newConnection, 4)";}
"(BulkVerifyVolumes, 1)" -> {"(setNodeVolume, 4)";}
"(syncPod, 1)" -> {"(SyncPod, 4)";}
"(Size, 0)" -> {"(GetLengthSid, 1)";}
"(refreshMuxLocked, 0)" -> {"(byPrefixPriority, 1)";}
"(Error, 3)" -> {"(errorS, 5)";}
"(singleRetrieve, 1)" -> {"(Retrieve, 0)";}
"(Dispatch, 4)" -> {}
"(RemoveInstance, 4)" -> {"(AddInstances, 4)";}
"(newWatchableStore, 6)" -> {"(syncVictimsLoop, 0)";}
"(ReadUvarint, 1)" -> {}
"(copyStdin, 0)" -> {}
"(Sum, 1)" -> {"(RawVector, 0)";}
"(SecondsSinceSSHKeySync, 0)" -> {"(addLivezChecks, 2)";}
"(PortForward, 3)" -> {"(portForwardFunc, 3)";}
"(wrapUp, 2)" -> {"(checkForUpdates, 1)";}
"(New, 1)" -> {"(newLoaderAtConfirmedDir, 4)";}
"(Terminate, 1)" -> {}
"(monitor, 0)" -> {}
"(serveConn, 1)" -> {}
"(RediscoverNode, 1)" -> {"(DiscoverNode, 1)";}
"(Flunders, 0)" -> {}
"(ReadAll, 1)" -> {"(readAll, 2)";}
"(constructBlock, 1)" -> {"(numLineBreaks, 1)";}
"(Ctx, 0)" -> {"(Delete, 2)";"(waitDeletes, 4)";}
"(updateStatusInternal, 3)" -> {"(updateLastTransitionTime, 3)";}
"(Put, 1)" -> {"(reset, 1)";"(reset, 2)";}
"(RemoteCredProvider, 2)" -> {"(executeCredentialProcess, 0)";"(createDialer, 0)";"(CreateHealthCheck, 1)";}
"(passesVerifyFilter, 2)" -> {"(verify, 1)";}
"(sync, 0)" -> {"(Observe, 1)";}
"(fn, 1)" -> {}
"(Write, 1)" -> {"(Send, 1)";"(printWithFileLine, 7)";}
"(Format, 1)" -> {"(ProtoReflect, 0)";}
"(ImportLayer, 4)" -> {"(layerPath, 2)";}
"(RestartNode, 1)" -> {}
"(evictOldest, 1)" -> {"(newWatcherGrpcStream, 1)";}
"(DeleteCollection, 4)" -> {}
"(NewClientTransport, 7)" -> {"(newHTTP2Client, 7)";}
"(ReduceFunc, 1)" -> {"(Offset, 1)";"(Id, 1)";}
"(TimeFromUnixNano, 1)" -> {}
"(isYesD, 0)" -> {"(NewClientConnectionWithPings, 2)";"(dispatch, 0)";}
"(newCollector, 2)" -> {"(mapEventsToCustomEvents, 1)";}
"(createNodeSecurityGroup, 5)" -> {"(toRuleProtocol, 1)";}
"(commit, 1)" -> {"(unsafeCommit, 1)";}
"(evict, 0)" -> {"(evictOldest, 1)";}
"(Destroy, 0)" -> {"(GetPath, 0)";}
"(timeoutIO, 0)" -> {"(configureTransports, 1)";"(enable, 0)";"(RoundTripOpt, 2)";"(StartRequest, 7)";"(resolveNow, 1)";"(NewClientConnectionWithPings, 2)";"(roundTrip, 1)";"(applyServiceConfigAndBalancer, 2)";"(Gather, 0)";"(newWatcherGrpcStream, 1)";"(dispatch, 0)";}
"(getDelegate, 0)" -> {"(NewDiscoveryRESTMapper, 1)";}
"(HandlePodUpdates, 1)" -> {"(dispatchWork, 4)";}
"(Delims, 2)" -> {"(roundTrip, 1)";"(RoundTripOpt, 2)";}
"(DecodeRuneInString, 1)" -> {"(errorf, 3)";"(GetUnknown, 0)";"(SetError, 0)";}
"(getTrimmed, 1)" -> {"(timeoutIO, 0)";}
"(quoteChar, 1)" -> {"(Dispatch, 4)";}
"(keysFromRemote, 1)" -> {}
"(Send, 1)" -> {"(SendMsg, 1)";}
"(Inf, 1)" -> {"(closeRemoteChannels, 0)";}
"(Revoke, 2)" -> {"(Values, 0)";}
"(pingTimeoutSeconds, 0)" -> {"(configureTransports, 1)";}
"(dispatchWork, 4)" -> {"(TerminatePod, 1)";}
"(copyStdout, 1)" -> {}
"(unmarshalPointer, 4)" -> {"(initializeCSINode, 1)";}
"(GetDetail, 0)" -> {"(CancelRequest, 1)";}
"(NewClientConn, 3)" -> {"(newMux, 1)";}
"(SetUpAt, 2)" -> {"(totalSecretBytes, 1)";}
"(marshalDoc, 2)" -> {"(nodev, 1)";}
"(NewLeaseFromLeaseClient, 3)" -> {"(WithCancel, 1)";}
"(GetWithContext, 1)" -> {"(singleRetrieve, 1)";}
"(SpecDidComplete, 1)" -> {"(failureDetails, 1)";}
"(ListenAndServeKubeletReadOnlyServer, 5)" -> {"(ListenAndServe, 0)";}
"(create, 3)" -> {"(logName, 2)";}
"(keepAliveLoop, 2)" -> {"(replyToClient, 2)";}
"(parse, 1)" -> {"(NewKey, 2)";}
"(WithLabelValues, 1)" -> {"(GetMetricWithLabelValues, 1)";}
"(log, 2)" -> {"(write, 0)";}
"(Build, 3)" -> {}
"(cadvisorInfoToNetworkStats, 1)" -> {"(latestContainerStats, 1)";}
"(newHTTP2Transport, 2)" -> {"(Warningln, 2)";}
"(Exit, 1)" -> {"(print, 4)";}
"(MapIndex, 1)" -> {"(equalMessage, 2)";}
"(AppendUint, 1)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(ToFloat64, 1)" -> {"(ToFloat64E, 1)";}
"(flushAll, 0)" -> {"(Sync, 0)";}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 0a839c6c3b9489716e216b4fc9dd2f9c17948c16
| 1.0 | kubernetes/kubernetes: pkg/controller/endpointslice/reconciler_test.go; 45 LoC -
Found a possible issue in [kubernetes/kubernetes](https://www.github.com/kubernetes/kubernetes) at [pkg/controller/endpointslice/reconciler_test.go](https://github.com/kubernetes/kubernetes/blob/0a839c6c3b9489716e216b4fc9dd2f9c17948c16/pkg/controller/endpointslice/reconciler_test.go#L444-L488)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to slice at line 476 may start a goroutine
[Click here to see the code in its original context.](https://github.com/kubernetes/kubernetes/blob/0a839c6c3b9489716e216b4fc9dd2f9c17948c16/pkg/controller/endpointslice/reconciler_test.go#L444-L488)
<details>
<summary>Click here to show the 45 line(s) of Go which triggered the analyzer.</summary>
```go
for _, slice := range slices {
if !strings.HasPrefix(slice.Name, testCase.service.Name) {
t.Fatalf("Expected EndpointSlice name to start with %s, got %s", testCase.service.Name, slice.Name)
}
if !reflect.DeepEqual(testCase.expectedLabels, slice.Labels) {
t.Errorf("Expected EndpointSlice to have labels: %v , got %v", testCase.expectedLabels, slice.Labels)
}
if slice.Labels[discovery.LabelServiceName] != testCase.service.Name {
t.Fatalf("Expected EndpointSlice to have label set with %s value, got %s", testCase.service.Name, slice.Labels[discovery.LabelServiceName])
}
if slice.Annotations[corev1.EndpointsLastChangeTriggerTime] != triggerTime.Format(time.RFC3339Nano) {
t.Fatalf("Expected EndpointSlice trigger time annotation to be %s, got %s", triggerTime.Format(time.RFC3339Nano), slice.Annotations[corev1.EndpointsLastChangeTriggerTime])
}
// validate that this slice has address type matching expected
expectedEndPointList := testCase.expectedEndpointPerSlice[slice.AddressType]
if expectedEndPointList == nil {
t.Fatalf("address type %v is not expected", slice.AddressType)
}
if len(slice.Endpoints) != len(expectedEndPointList) {
t.Fatalf("Expected %v Endpoint, got %d", len(expectedEndPointList), len(slice.Endpoints))
}
// test is limited to *ONE* endpoint
endpoint := slice.Endpoints[0]
if !reflect.DeepEqual(endpoint, expectedEndPointList[0]) {
t.Fatalf("Expected endpoint: %+v, got: %+v", expectedEndPointList[0], endpoint)
}
expectTrackedResourceVersion(t, r.endpointSliceTracker, &slice, "100")
expectMetrics(t,
expectedMetrics{
desiredSlices: 1,
actualSlices: 1,
desiredEndpoints: 1,
addedPerSync: len(testCase.expectedEndpointPerSlice),
removedPerSync: 0,
numCreated: len(testCase.expectedEndpointPerSlice),
numUpdated: 0,
numDeleted: 0})
}
```
</details>
<details>
<summary>Click here to show extra information the analyzer produced.</summary>
```
The following dot graph describes paths through the callgraph that could lead to a function calling a goroutine:
digraph G {
"(writeback, 1)" -> {"(merge, 1)";}
"(scheduleBodyWrite, 0)" -> {}
"(shutDownIn, 1)" -> {}
"(sendPings, 1)" -> {}
"(WithTimeout, 2)" -> {"(WithDeadline, 2)";}
"(Example, 0)" -> {}
"(RpcError_ErrorCode, 1)" -> {"(CancelRequest, 1)";}
"(RetrieveOne, 4)" -> {"(sendConsumeCustomMetric, 5)";"(ServerPreferredResources, 1)";"(RediscoverNode, 1)";"(Run, 4)";"(sendLoop, 0)";"(SetUpAt, 2)";"(newConn, 2)";"(createMergedSnapshotMessage, 4)";"(createAndInitKubelet, 30)";"(NewClientConn, 3)";"(Evaluate, 1)";"(SpecDidComplete, 1)";"(ListenAndServeReadOnly, 3)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(on100, 0)";"(RestartNode, 1)";"(mountServiceAccountToken, 2)";"(NewSession, 2)";"(HandlePodReconcile, 1)";"(DeleteCollection, 4)";"(DiscoverNode, 1)";"(sendConsumeMemRequests, 4)";"(ConfigureTransports, 1)";"(BeforeSuiteDidRun, 1)";"(rollbackEtcd3MinorVersion, 2)";"(newETCD3Storage, 2)";"(keepAliveLoop, 2)";"(SpecSuiteWillBegin, 2)";"(SpecSuiteDidEnd, 1)";"(apply, 3)";"(Dispatch, 4)";"(handleResponse, 2)";"(recvLoop, 0)";"(Build, 3)";"(processPodOnNode, 5)";"(Sink, 0)";"(NewDelayingQueueWithCustomClock, 2)";"(RoundTripOpt, 2)";"(CloseNotify, 0)";"(startCompactorOnce, 2)";"(GetCollector, 1)";"(Chan, 0)";"(addPeerToProber, 6)";"(NewClientConnectionWithPings, 2)";"(PortForward, 3)";"(createVolume, 2)";"(NewProxier, 9)";"(NewStreamWatcher, 2)";"(Eventf, 7)";"(listenerRoutine, 0)";"(VerifySignature, 2)";"(UpdatePod, 1)";"(NewMainKubelet, 30)";"(sendMergedSnap, 1)";"(receive, 0)";"(StartTestServer, 4)";"(dispatch, 0)";"(HandleStreams, 2)";"(mergeService, 1)";"(Dispatch, 3)";"(runSTM, 2)";"(StartGarbageCollection, 0)";"(sendUpdates, 0)";"(roundTrip, 1)";"(UpdateTransport, 4)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(updateTransport, 5)";"(markDiskAsDeletable, 2)";"(applySnapshot, 2)";"(enable, 0)";"(recordToSink, 2)";"(newLessor, 3)";"(HandlePodCleanups, 0)";"(RunOnce, 1)";"(PortForward, 4)";"(getMetricsFromNode, 2)";"(NewAggregator, 4)";"(NewProxier, 13)";"(newClientConn, 2)";"(runOnce, 2)";"(AttachVolume, 2)";"(dialStreamLocal, 1)";"(AddPod, 1)";"(newSnapshotReaderCloser, 2)";"(sendConsumeCPURequests, 4)";"(AfterSuiteDidRun, 1)";"(newFilePipeline, 3)";"(startFrameWrite, 1)";"(BuildAuthn, 2)";"(newMux, 1)";"(dgemmParallel, 12)";"(SendSnapshot, 1)";"(monitorSize, 2)";"(podsRunning, 2)";"(GenerateRegisterPluginFunc, 4)";"(applyEntries, 2)";"(newWatcherGrpcStream, 1)";"(HandlePodUpdates, 1)";"(CreatePD, 1)";"(verify, 1)";"(newDecoratedWatcher, 2)";"(handleConnection, 2)";"(Launch, 1)";"(scheduleFrameWrite, 0)";"(handleSignal, 1)";"(removeAndReAdd, 1)";"(callK, 1)";"(newProxyServer, 3)";"(ListenAndServe, 8)";"(SyncServices, 1)";"(StartInterceptingOutput, 0)";"(createCall, 5)";"(CreateVolume, 3)";"(newHTTP2Transport, 2)";"(Verify, 2)";"(startKubelet, 6)";"(StartNode, 2)";"(startPeer, 4)";"(waitCancelSubstreams, 1)";"(initializeCSINode, 1)";"(startDBSizeMonitorPerEndpoint, 2)";"(clientHandshake, 2)";}
"(UpdateObjectCount, 2)" -> {"(newTimeBudget, 1)";}
"(AddHTTP, 3)" -> {}
"(Search, 2)" -> {"(Execute, 2)";}
"(FromContext, 1)" -> {"(Unknown, 1)";}
"(callTRFunc, 3)" -> {}
"(replyToClient, 2)" -> {"(deadlineLoop, 0)";}
"(runOnce, 2)" -> {"(getFailedContainers, 1)";}
"(startCaching, 1)" -> {}
"(convertProcessError, 2)" -> {"(processSignalResult, 2)";}
"(processData, 0)" -> {"(onEachWrite, 0)";}
"(syncVictimsLoop, 0)" -> {"(moveVictims, 0)";}
"(cleanupOrphanedPodCgroups, 3)" -> {}
"(InstallAPIs, 3)" -> {"(InstallAPIGroups, 1)";}
"(NewTicker, 1)" -> {"(yaml_emitter_state_machine, 2)";}
"(newClientV3, 1)" -> {"(toGRPC, 1)";}
"(pathsToRemove, 2)" -> {"(Difference, 1)";}
"(LastIndex, 0)" -> {"(passesVerifyFilter, 2)";"(initializeCSINode, 1)";"(dispatch, 0)";"(runSTM, 2)";"(freepages, 0)";"(newETCD3Storage, 2)";"(NewClientConnectionWithPings, 2)";"(roundTrip, 1)";"(NormalizedName, 1)";"(AddPod, 1)";"(RoundTripOpt, 2)";}
"(TokenVolumeSource, 0)" -> {}
"(newAffinityTerm, 2)" -> {"(GetNamespacesFromPodAffinityTerm, 2)";}
"(waitDelete, 4)" -> {"(Watch, 3)";"(freepages, 0)";}
"(wrapErr, 1)" -> {}
"(NewMainKubelet, 30)" -> {"(defaultNodeStatusFuncs, 0)";}
"(serveUDP, 1)" -> {"(spawnWorker, 1)";}
"(string, 1)" -> {"(DecodeRuneInString, 1)";}
"(Encode, 2)" -> {"(msgHeader, 2)";}
"(NewClusterV3, 2)" -> {"(NewClientV3, 1)";}
"(legacyLogSymlink, 4)" -> {"(logSymlink, 4)";}
"(getAgentPoolScaleSets, 1)" -> {"(CreateOrUpdateVMSS, 3)";}
"(msgHeader, 2)" -> {"(Do, 2)";"(wrapErr, 1)";}
"(handleConnection, 2)" -> {}
"(Exec, 2)" -> {"(AbsHostPath, 1)";}
"(RemoveImage, 1)" -> {"(RemoveImage, 2)";}
"(standardMethodArgumentDecode, 4)" -> {}
"(logSymlink, 4)" -> {"(transformStartContainerError, 1)";"(CreateComputeSystem, 3)";}
"(SpecSuiteDidEnd, 1)" -> {"(AnnounceSpecRunCompletion, 2)";}
"(From, 1)" -> {"(To, 1)";}
"(ToJSONStream, 1)" -> {"(Chan, 0)";"(on100, 0)";"(VerifySignature, 2)";"(serve, 0)";"(lockedUpdateLoadBalancerHosts, 2)";"(startFrameWrite, 1)";"(CloseNotify, 0)";"(createVolume, 2)";"(NewTimeoutListener, 5)";"(Eventf, 7)";"(newSnapshotReaderCloser, 2)";"(PortForward, 3)";"(SendSnapshot, 1)";"(Dispatch, 4)";}
"(OrNil, 0)" -> {}
"(add, 1)" -> {"(evict, 0)";}
"(Version, 0)" -> {"(NewVersion, 1)";}
"(, 2)" -> {"(getRequest, 1)";}
"(CreateHealthCheck, 1)" -> {"(HealthChecks, 0)";"(newETCD3HealthCheck, 1)";}
"(toGRPC, 1)" -> {"(ElectionServerToElectionClient, 1)";}
"(GRPCStatus, 0)" -> {"(Exp, 1)";"(SetTypeForKey, 2)";"(uint64, 0)";"(getFileType, 1)";}
"(erorr, 0)" -> {"(processSignalResult, 2)";}
"(startPeer, 4)" -> {"(newSnapshotSender, 4)";}
"(defaultNodeStatusFuncs, 0)" -> {"(VolumesInUse, 2)";}
"(NewProxier, 9)" -> {"(NewCustomProxier, 10)";}
"(layerPath, 2)" -> {}
"(NewOpenIDMetadata, 4)" -> {"(openIDKeysetJSON, 1)";}
"(Sync, 0)" -> {"(errnoErr, 1)";}
"(newCluster, 2)" -> {"(fillClusterForMembers, 0)";}
"(IterateUsing, 2)" -> {"(eachMapEntry, 2)";}
"(removeBlock, 1)" -> {"(Clear, 0)";"(paragraph, 1)";}
"(Clone, 0)" -> {"(NewTLS, 1)";}
"(Evaluate, 1)" -> {}
"(waitForChanges, 0)" -> {"(ChangeEvents, 2)";}
"(Sign, 0)" -> {"(Run, 1)";}
"(Head, 2)" -> {"(keyIndex, 1)";}
"(configureHTTP2Transport, 1)" -> {"(pingTimeoutSeconds, 0)";}
"(Config, 0)" -> {"(toStringSlice, 1)";}
"(New, 3)" -> {"(processFlag, 1)";}
"(hcsCloseProcess, 1)" -> {}
"(SortStrings, 1)" -> {}
"(topIndex, 0)" -> {"(Dispatch, 4)";}
"(max, 1)" -> {}
"(updateResolverState, 2)" -> {"(updateClientConnState, 1)";}
"(numLineBreaks, 1)" -> {"(newWatcherGrpcStream, 1)";}
"(serveListen, 0)" -> {"(serveConn, 1)";}
"(monitorSize, 2)" -> {"(monitorResizeEvents, 3)";}
"(empty, 0)" -> {"(StringAffineComparable, 1)";}
"(dispatch, 0)" -> {}
"(Eventf, 7)" -> {}
"(handleReadLoopResponse, 1)" -> {}
"(append, 1)" -> {}
"(flushColdBuf, 0)" -> {"(maybeRotateStreams, 0)";}
"(sendLoop, 0)" -> {"(RequestProgress, 1)";}
"(OpenComputeSystem, 2)" -> {}
"(GetClientConn, 2)" -> {"(getClientConn, 3)";}
"(TunnelSyncHealthChecker, 1)" -> {"(SecondsSinceSSHKeySync, 0)";}
"(Done, 0)" -> {"(erorr, 0)";}
"(CloseNotify, 0)" -> {}
"(switchBalancer, 1)" -> {"(newCCBalancerWrapper, 3)";}
"(getFailedContainers, 1)" -> {"(syncPod, 1)";}
"(processFlag, 1)" -> {"(setState, 1)";}
"(ErrAllFailed, 1)" -> {"(verify, 1)";}
"(StartGatheringData, 0)" -> {}
"(RemoveImage, 2)" -> {"(ImageRemove, 3)";}
"(Zcopy, 5)" -> {"(BrowseDiagnosticLog, 3)";}
"(hive, 1)" -> {"(StartGarbageCollection, 0)";"(sendUpdates, 0)";"(NewProxier, 9)";"(ListenAndServe, 8)";"(ListenAndServeReadOnly, 3)";"(NewMainKubelet, 30)";"(runOnce, 2)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(StartInterceptingOutput, 0)";}
"(cancel, 2)" -> {"(removeChild, 2)";}
"(NewCacherFromConfig, 1)" -> {"(startCaching, 1)";}
"(Keys, 1)" -> {"(KeysPrefix, 2)";}
"(freeTrace, 1)" -> {"(provider, 0)";}
"(resetTransport, 0)" -> {"(startHealthCheck, 1)";}
"(nodev, 1)" -> {"(node, 2)";}
"(Launch, 1)" -> {"(waitVersion, 0)";}
"(walkParameters, 1)" -> {"(HandleStreams, 2)";"(StartRequest, 7)";"(resolveNow, 1)";"(NewServerTransport, 3)";"(applyServiceConfigAndBalancer, 2)";"(NewClusterV3, 2)";"(makeConnectedServerPipe, 0)";}
"(DefaultCollector, 1)" -> {"(newConn, 2)";"(dialStreamLocal, 1)";"(createCall, 5)";}
"(startStreamWriter, 6)" -> {}
"(GetContainerInfoV2, 2)" -> {"(OrNil, 0)";}
"(GetRequestedContainersInfo, 2)" -> {"(OrNil, 0)";}
"(effectivelyEquals, 1)" -> {"(IsSelected, 1)";}
"(PollImmediateUntil, 3)" -> {"(PollUntil, 3)";}
"(Remove, 1)" -> {"(removeBlock, 1)";}
"(errorf, 3)" -> {"(CreateHealthCheck, 1)";"(executeCredentialProcess, 0)";"(createCall, 5)";"(ConfigureTransports, 1)";"(RediscoverNode, 1)";"(createDialer, 0)";}
"(executeCredentialProcess, 0)" -> {}
"(reflectValue, 1)" -> {"(valueEncoder, 1)";}
"(GetLatestPage, 0)" -> {"(RawSyscallNoError, 4)";}
"(SendRequest, 3)" -> {"(fn, 1)";}
"(Get, 1)" -> {"(GetUsing, 2)";}
"(createAndInitKubelet, 30)" -> {"(StartGarbageCollection, 0)";}
"(NewTLS, 1)" -> {"(AddHTTP, 3)";"(mustNewMember, 1)";"(Launch, 0)";"(startStreamWriter, 6)";}
"(deleteContainersInPod, 3)" -> {"(getContainersToDeleteInPod, 3)";}
"(NewClientHandler, 3)" -> {"(requestLogger, 2)";}
"(CreateOrUpdateVMSS, 3)" -> {"(PutResources, 2)";}
"(update, 4)" -> {"(CanSet, 0)";}
"(yaml_parser_set_input_string, 2)" -> {}
"(debugResponse, 1)" -> {"(newTeeReader, 2)";}
"(mkSTM, 2)" -> {"(runSTM, 2)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(GetClientConn, 2)";}
"(write, 0)" -> {"(fdatasync, 1)";}
"(newFilePipeline, 3)" -> {}
"(newCCBalancerWrapper, 3)" -> {"(Build, 2)";}
"(handler, 0)" -> {"(Lstat, 1)";}
"(formatUnequalValues, 2)" -> {"(truncatingFormat, 1)";}
"(V, 1)" -> {"(setV, 1)";}
"(GetLengthSid, 1)" -> {"(GetFileType, 1)";}
"(convertSliceNumbers, 2)" -> {"(convertNumber, 1)";}
"(runHandler, 1)" -> {"(handler, 0)";}
"(Do, 1)" -> {"(ReadAll, 1)";}
"(parseProposeCtxErr, 2)" -> {"(Upload, 4)";}
"(GenerateBulkVolumeVerifyFunc, 4)" -> {"(OperationCompleteHook, 2)";"(GetFullQualifiedPluginNameForVolume, 2)";}
"(shutdown, 1)" -> {}
"(copyStderr, 1)" -> {}
"(getFloatingIPByPortID, 2)" -> {"(selectLoadBalancer, 4)";}
"(Set, 1)" -> {"(isLiteral, 1)";}
"(checkTxnAuth, 3)" -> {"(checkTxnReqsPermission, 3)";}
"(StartTestServerOrDie, 4)" -> {"(StartTestServer, 4)";}
"(SetEndpoints, 1)" -> {"(NewAddress, 1)";}
"(kind, 0)" -> {"(uint64, 0)";}
"(ok, 0)" -> {"(satisfiedExpectations, 0)";}
"(recvLoop, 0)" -> {}
"(isControl, 0)" -> {}
"(Add, 2)" -> {"(SetScale, 1)";}
"(Signal, 1)" -> {"(Marshal, 1)";}
"(DeleteProducer, 1)" -> {}
"(Next, 0)" -> {"(lookup, 0)";}
"(NsHandle, 1)" -> {"(callopen64, 3)";}
"(HealthChecks, 0)" -> {}
"(listenerRoutine, 0)" -> {"(makeConnectedServerPipe, 0)";}
"(Do, 0)" -> {}
"(printDepth, 5)" -> {"(Filter, 1)";"(output, 6)";"(Bytes, 0)";"(getBuffer, 0)";}
"(sendConsumeMemRequests, 4)" -> {}
"(getRequestedContainers, 2)" -> {}
"(executePreStopHook, 4)" -> {}
"(Properties, 4)" -> {"(DefaultCollector, 1)";}
"(readAll, 2)" -> {"(SetUnknown, 1)";}
"(autoSync, 0)" -> {"(Lvl, 1)";}
"(Get, 4)" -> {"(NewError, 3)";}
"(Insert, 1)" -> {"(Descend, 1)";}
"(AddPod, 1)" -> {"(VisitPodConfigmapNames, 2)";}
"(newBpsLoop, 1)" -> {}
"(newBackend, 1)" -> {"(run, 0)";}
"(AddExtensionType, 2)" -> {"(getTrimmed, 1)";"(NewTicker, 1)";"(yaml_emitter_emit, 2)";}
"(Unlock, 1)" -> {"(Delete, 2)";}
"(getWrappedWorkerFunc, 1)" -> {}
"(NewKey, 2)" -> {}
"(createMetadatas, 1)" -> {"(WithTTL, 1)";}
"(AppendUint64, 1)" -> {"(AppendUint, 1)";}
"(apply, 3)" -> {"(AddNamedImport, 4)";}
"(GetCgroupStats, 2)" -> {"(cadvisorInfoToNetworkStats, 1)";}
"(rollbackEtcd3MinorVersion, 2)" -> {"(Snapshot, 2)";}
"(newTLSListener, 3)" -> {"(acceptLoop, 0)";}
"(Transform, 3)" -> {"(kind, 0)";}
"(maybeRotateStreams, 0)" -> {"(Descend, 1)";}
"(NewConnection, 2)" -> {"(monitor, 0)";}
"(previousDestroy, 0)" -> {"(UpdateObjectCount, 2)";}
"(waitCancelSubstreams, 1)" -> {}
"(listPodStats, 1)" -> {"(makePodStorageStats, 2)";}
"(output, 6)" -> {"(timeoutFlush, 1)";"(int64, 1)";"(Format, 1)";"(Reset, 0)";"(Error, 0)";}
"(Dispatch, 3)" -> {"(Dispatch, 4)";}
"(NewClientV3, 1)" -> {"(newClientV3, 1)";}
"(provider, 0)" -> {"(newWatcherGrpcStream, 1)";"(Dispatch, 4)";}
"(ListPodStats, 0)" -> {"(listPodStats, 1)";}
"(DialWithConn, 4)" -> {"(newMux, 1)";"(HandleStreams, 2)";"(AttachVolume, 2)";"(Chan, 0)";"(removeAndReAdd, 1)";"(DeleteCollection, 4)";"(startFrameWrite, 1)";"(Dispatch, 4)";"(NewMainKubelet, 30)";"(sendUpdates, 0)";"(rollbackEtcd3MinorVersion, 2)";"(HandlePodCleanups, 0)";"(GenerateRegisterPluginFunc, 4)";"(runSTM, 2)";"(SyncServices, 1)";"(StartInterceptingOutput, 0)";"(StartTestServer, 4)";"(CloseNotify, 0)";"(sendConsumeCustomMetric, 5)";"(CreatePD, 1)";"(startDBSizeMonitorPerEndpoint, 2)";"(Launch, 0)";"(mergeService, 1)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(dispatch, 0)";"(sendConsumeMemRequests, 4)";"(NewProxier, 9)";"(handleConnection, 2)";"(CreateVolume, 3)";"(newSnapshotReaderCloser, 2)";"(dgemmParallel, 12)";"(updateTransport, 5)";"(NewDelayingQueueWithCustomClock, 2)";"(ListenAndServe, 8)";"(sendConsumeCPURequests, 4)";"(ListenAndServeReadOnly, 3)";"(Eventf, 7)";"(PortForward, 3)";"(createVolume, 2)";"(VerifySignature, 2)";"(NewClientConnectionWithPings, 2)";"(startCompactorOnce, 2)";"(SendSnapshot, 1)";"(processPodOnNode, 5)";"(newWatcherGrpcStream, 1)";"(recordToSink, 2)";"(Run, 4)";"(UpdatePod, 1)";"(on100, 0)";"(monitorResizeEvents, 3)";"(DiscoverNode, 1)";"(clientHandshake, 2)";"(StartGarbageCollection, 0)";"(runOnce, 2)";"(handleSignal, 1)";"(verify, 1)";}
"(fetchGroupVersionResources, 2)" -> {}
"(eachMapEntry, 2)" -> {"(dialStreamLocal, 1)";"(Launch, 1)";"(Build, 3)";"(createAndInitKubelet, 30)";"(UpdateTransport, 4)";"(newFilePipeline, 3)";"(newSession, 2)";"(BuildAuthn, 2)";"(StartNode, 2)";"(StartTestServerOrDie, 4)";"(callK, 1)";"(MigrateIfNeeded, 1)";"(inWorker, 0)";"(Sink, 0)";"(receive, 0)";"(addPeerToProber, 6)";"(mountServiceAccountToken, 2)";"(newConn, 2)";"(NewProxier, 13)";"(SetUpAt, 2)";"(RediscoverNode, 1)";"(Verify, 2)";"(copyStdin, 0)";"(StartGatheringData, 0)";"(enable, 0)";"(ServerGroupsAndResources, 1)";"(NewAggregator, 4)";"(waitForAvailableEndpoint, 1)";"(newClientConn, 2)";"(roundTrip, 1)";"(handleResizes, 0)";"(Evaluate, 1)";"(sendLoop, 0)";"(RoundTripOpt, 2)";"(Dispatch, 3)";"(createCall, 5)";"(freepages, 0)";"(copyStderr, 1)";"(initializeCSINode, 1)";"(CopyAllLogs, 4)";"(newETCD3Storage, 2)";"(NewClientConn, 3)";"(Dgemm, 13)";"(applySnapshot, 2)";"(copyStdout, 1)";"(createMergedSnapshotMessage, 4)";"(scheduleFrameWrite, 0)";"(passesVerifyFilter, 2)";"(ListenForSignal, 1)";"(recvLoop, 0)";"(NewSession, 2)";"(RunOnce, 1)";"(listenerRoutine, 0)";"(newHTTP2Transport, 2)";"(newProxyServer, 3)";"(watchErrorStream, 2)";"(sendMergedSnap, 1)";"(podsRunning, 2)";"(handleResponse, 2)";"(SpecSuiteWillBegin, 2)";"(SpecSuiteDidEnd, 1)";"(PortForward, 4)";"(waitCancelSubstreams, 1)";"(applyEntries, 2)";"(startKubelet, 6)";"(startPeer, 4)";"(createPD, 1)";"(RestartNode, 1)";"(newDecoratedWatcher, 2)";"(AddPod, 1)";"(getMetricsFromNode, 2)";"(WatchPods, 4)";"(GetCollector, 1)";"(NewStreamWatcher, 2)";"(monitorSize, 2)";"(DefineTests, 2)";"(newLessor, 3)";}
"(CreateComputeSystem, 3)" -> {}
"(Values, 0)" -> {"(retrieveUnexportedField, 3)";}
"(yaml_emitter_emit, 2)" -> {"(handleInterrupt, 0)";"(ConfigureTransports, 1)";"(callK, 1)";}
"(newTeeReader, 2)" -> {"(newWatcherGrpcStream, 1)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(Connect, 0)";"(GetClientConn, 2)";}
"(Flush, 0)" -> {"(Split, 2)";"(lockAndFlushAll, 0)";}
"(Split, 2)" -> {"(writeDataFromHandler, 3)";}
"(PortForward, 4)" -> {"(PortForward, 3)";}
"(insert, 2)" -> {}
"(NewNamingConditionController, 2)" -> {"(AddEventHandler, 1)";}
"(sweetenFields, 1)" -> {"(append, 1)";"(Remove, 1)";"(pathsToRemove, 2)";"(Wait, 0)";"(openExistingOrNew, 1)";"(Header, 0)";"(Decode, 1)";"(Do, 1)";"(V, 1)";"(Transform, 3)";"(Sync, 0)";"(Call, 5)";"(Flush, 0)";"(Index, 2)";"(flushColdBuf, 0)";"(New, 3)";"(Add, 2)";"(Parse, 1)";"(New, 0)";"(Update, 3)";"(Add, 1)";"(Send, 1)";"(Put, 1)";"(Copy, 2)";"(rotate, 0)";"(Encode, 2)";"(Reset, 0)";}
"(NewController, 1)" -> {"(newCallSet, 0)";}
"(Client, 1)" -> {"(NewClient, 2)";}
"(dedup, 1)" -> {"(LoadSources, 3)";}
"(Infof, 2)" -> {"(Printf, 2)";}
"(sendKeepAliveLoop, 1)" -> {"(LeaseKeepAlive, 1)";}
"(onEachWrite, 0)" -> {"(SendSnapshot, 1)";"(newSnapshotReaderCloser, 2)";"(CloseNotify, 0)";"(on100, 0)";"(startFrameWrite, 1)";"(PortForward, 3)";}
"(delete, 1)" -> {"(empty, 0)";}
"(testScriptInPod, 4)" -> {"(GetTestImage, 1)";}
"(GetUsing, 2)" -> {"(findJsonNameField, 1)";}
"(outPayload, 5)" -> {"(finishStream, 5)";"(sendConsumeCPURequests, 4)";"(CloseNotify, 0)";"(DeleteCollection, 4)";"(sendConsumeCustomMetric, 5)";"(runSTM, 2)";"(GenerateBulkVolumeVerifyFunc, 4)";"(NewClientConnectionWithPings, 2)";"(Run, 4)";"(newWatcherGrpcStream, 1)";"(recordToSink, 2)";"(addEvent, 3)";"(NewDelayingQueueWithCustomClock, 2)";"(RoundTripOpt, 2)";"(dispatch, 0)";"(sendConsumeMemRequests, 4)";"(HandleStreams, 2)";"(roundTrip, 1)";"(markDiskAsDeletable, 2)";}
"(Quo, 2)" -> {"(GetClientConn, 2)";"(dispatch, 0)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(NewClientConnectionWithPings, 2)";}
"(StartNode, 2)" -> {}
"(NewSession, 2)" -> {"(KeepAlive, 2)";}
"(createVolume, 2)" -> {"(TokenVolumeSource, 0)";}
"(Load, 0)" -> {"(Start, 1)";"(Get, 1)";"(Next, 0)";"(PollImmediateUntil, 3)";"(NewDiscoveryController, 3)";"(stopServers, 2)";"(newBackend, 1)";"(sync, 0)";"(Encode, 2)";"(NewREST, 2)";"(commit, 1)";"(Add, 2)";"(NewWithTLSConfig, 2)";"(Read, 1)";"(SetEndpoints, 1)";"(Node, 0)";"(getDelegate, 0)";"(ListPodStatsAndUpdateCPUNanoCoreUsage, 0)";"(From, 1)";"(keyToFilename, 1)";"(Start, 0)";"(Unmarshal, 2)";"(HcsCloseProcess, 2)";"(Set, 1)";"(Handle, 2)";"(New, 3)";"(, 0)";"(Terminate, 0)";"(V1, 0)";"(Verifier, 1)";"(Get, 4)";"(Write, 1)";"(ImportLayer, 4)";"(Search, 2)";"(remove, 1)";"(WithTimeout, 2)";"(ListPodStats, 0)";"(NewNamingConditionController, 2)";"(New, 1)";"(Initialize, 2)";"(ServiceTicket, 3)";"(ok, 0)";"(NewCustomResourceDefinitionHandler, 15)";"(get, 1)";"(SendRequest, 3)";"(close, 0)";"(Lock, 1)";"(NewCRDFinalizer, 3)";"(shutdown, 1)";"(NewEstablishingController, 2)";"(reset, 0)";"(Config, 0)";"(newClient, 1)";"(Get, 3)";"(Get, 2)";"(delete, 1)";"(Sub, 1)";"(V, 1)";"(enableCSM, 3)";"(NewDesc, 4)";"(InstallAPIs, 3)";"(fetch, 1)";"(Reset, 0)";"(Zero, 0)";"(NewConditionController, 2)";"(tearDown, 1)";"(Wait, 0)";"(Decode, 1)";"(GetFromPath, 1)";"(Do, 0)";"(ToFloat64, 1)";"(Keys, 1)";"(unregisterCallback, 1)";"(NewClient, 2)";"(add, 1)";"(Do, 1)";"(Remove, 1)";"(Apiextensions, 0)";"(PollUntil, 3)";"(Run, 3)";"(New, 0)";"(Delete, 1)";"(Run, 2)";"(Run, 1)";"(apply, 2)";"(Add, 1)";"(newStore, 1)";"(Transform, 3)";"(Install, 1)";"(New, 2)";"(parse, 1)";"(NewClusterAuthenticationTrustController, 2)";"(New, 6)";"(NewSessionWithOptions, 1)";"(Sign, 0)";"(Equal, 3)";"(Sync, 0)";"(Flush, 0)";"(NewOpenIDMetadata, 4)";"(GetCgroupStats, 2)";"(reset, 1)";"(Get, 0)";"(Client, 1)";"(Parse, 1)";"(Put, 1)";"(NewCtxClient, 1)";"(Size, 0)";"(init, 0)";"(HcsCloseComputeSystem, 2)";"(Send, 1)";"(NewKubernetesAPIApprovalPolicyConformantConditionController, 2)";"(Format, 1)";"(newAuthenticator, 2)";"(GetWithContext, 1)";"(Unlock, 1)";"(Revoke, 2)";"(NewLeaseFromLeaseClient, 3)";"(Error, 0)";"(ToJSON, 1)";"(installTunneler, 2)";"(NewController, 1)";}
"(poller, 2)" -> {}
"(awaitRequestCancel, 1)" -> {}
"(NewCustomProxier, 10)" -> {"(createProxier, 10)";}
"(ClockWait, 3)" -> {}
"(NewDefaultTableConvertor, 1)" -> {"(previousDestroy, 0)";}
"(Index, 2)" -> {"(Insert, 1)";}
"(Untranspose, 0)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(newWatcherGrpcStream, 1)";"(DeleteCollection, 4)";"(sendConsumeCPURequests, 4)";"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";}
"(on100, 0)" -> {}
"(UserDefinedMetricType, 1)" -> {"(getRequestedContainers, 2)";}
"(GarbageCollect, 0)" -> {"(DrainedAt, 0)";}
"(ListenAndServeKubeletServer, 10)" -> {"(ListenAndServe, 0)";}
"(ListenAndServePodResources, 0)" -> {"(ListenAndServePodResources, 4)";}
"(printWithFileLine, 7)" -> {"(create, 3)";}
"(allocate, 0)" -> {"(roundTrip, 1)";"(GetRequestedContainersInfo, 2)";"(fetchContainerStats, 1)";"(RoundTripOpt, 2)";"(callTTBFunc, 3)";"(apply, 3)";"(GetContainerInfoV2, 2)";}
"(totalSecretBytes, 1)" -> {}
"(GenerateRegisterPluginFunc, 4)" -> {"(RegisterPlugin, 3)";}
"(containerStartSpec, 1)" -> {"(legacyLogSymlink, 4)";}
"(printDelimiter, 0)" -> {"(ChangeEvents, 2)";}
"(UpdateLoadBalancer, 4)" -> {"(updateSecurityGroup, 4)";}
"(Copy, 2)" -> {"(Zcopy, 5)";}
"(StartGarbageCollection, 0)" -> {"(GarbageCollect, 0)";}
"(spawnWorker, 1)" -> {}
"(stopServers, 2)" -> {"(GracefulStop, 0)";}
"(String, 0)" -> {"(Iterate, 1)";}
"(makePodStorageStats, 2)" -> {"(cadvisorInfoToAcceleratorStats, 1)";}
"(newDelayingQueue, 3)" -> {}
"(processPodOnNode, 5)" -> {"(GetWorkerUnsafe, 1)";}
"(platformDoesNotSupportError, 1)" -> {"(isFeatureInRange, 2)";}
"(FatalDepth, 2)" -> {"(printDepth, 5)";}
"(handleResponse, 2)" -> {"(awaitRequestCancel, 1)";}
"(getBuffer, 0)" -> {"(Reset, 0)";}
"(initializeCSINode, 1)" -> {}
"(wrapTLS, 3)" -> {"(newTLSListener, 3)";}
"(expectTrackedResourceVersion, 4)" -> {"(Fatalf, 2)";"(Errorf, 4)";}
"(merge, 1)" -> {"(WithLazyPrefix, 1)";}
"(handleResizes, 0)" -> {}
"(RequestProgress, 1)" -> {}
"(Properties, 2)" -> {}
"(errnoErr, 1)" -> {"(StrategicMerge, 1)";"(msync, 1)";}
"(getFileType, 1)" -> {}
"(moveVictims, 0)" -> {"(newDelayingQueue, 3)";"(UpdateLoadBalancer, 4)";"(GetLoadBalancer, 3)";}
"(setV, 1)" -> {"(FileLine, 1)";}
"(compact, 3)" -> {"(Launch, 0)";"(NewProxier, 9)";"(ListenAndServe, 8)";"(recordToSink, 2)";"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(runOnce, 2)";"(monitorResizeEvents, 3)";"(sendUpdates, 0)";"(StartInterceptingOutput, 0)";"(newHTTP2Transport, 2)";"(ListenAndServeReadOnly, 3)";}
"(SetUnknown, 1)" -> {"(NewClientConnectionWithPings, 2)";"(markDiskAsDeletable, 2)";"(dispatch, 0)";}
"(callopen64, 3)" -> {"(openDir, 1)";}
"(newCond, 0)" -> {"(method, 0)";}
"(format, 2)" -> {"(format, 3)";}
"(Caller, 1)" -> {"(Next, 0)";}
"(ParseNetlinkMessage, 1)" -> {}
"(Printf, 2)" -> {"(output, 1)";}
"(isMembersEqual, 2)" -> {"(SortableMemberSliceByPeerURLs, 1)";}
"(GetCollector, 1)" -> {"(newCollector, 2)";}
"(visitContainerConfigmapNames, 2)" -> {"(newAffinityTerm, 2)";"(newIntSet, 0)";}
"(installTunneler, 2)" -> {"(TunnelSyncHealthChecker, 1)";}
"(storeScratch, 4)" -> {"(inBounds, 3)";}
"(Add, 1)" -> {"(ref, 0)";}
"(ToFloat64E, 1)" -> {"(indirect, 1)";}
"(Descend, 1)" -> {"(iterate, 6)";}
"(Sink, 0)" -> {}
"(shouldRetryRequest, 3)" -> {"(canRetryError, 1)";}
"(Filter, 1)" -> {"(Set, 2)";}
"(SetThumbprint, 2)" -> {}
"(AddEventHandler, 1)" -> {"(AddEventHandlerWithResyncPeriod, 2)";}
"(connect, 0)" -> {"(resetTransport, 0)";}
"(CreateDisk, 5)" -> {}
"(listenGRPC, 0)" -> {"(newBridge, 1)";}
"(ImageRemove, 3)" -> {"(executePreStopHook, 4)";}
"(get, 0)" -> {"(Load, 0)";}
"(GetPath, 0)" -> {"(GetIntelRdtPath, 1)";}
"(reset, 1)" -> {"(newReadOnly, 1)";}
"(HandleStreams, 2)" -> {}
"(waitVersion, 0)" -> {"(isMembersEqual, 2)";}
"(startHealthCheck, 1)" -> {"(healthCheckFunc, 4)";}
"(Decode, 1)" -> {"(newDecoder, 0)";}
"(Observe, 1)" -> {"(asyncFlush, 1)";}
"(startKubelet, 6)" -> {"(ListenAndServePodResources, 0)";}
"(Upload, 4)" -> {"(NewReader, 4)";}
"(decodeArguments, 3)" -> {"(standardMethodArgumentDecode, 4)";}
"(, 0)" -> {"(newReflectCache, 0)";}
"(Line, 0)" -> {"(lookup, 0)";}
"(finishStream, 5)" -> {"(addEvent, 3)";}
"(NewKubernetesAPIApprovalPolicyConformantConditionController, 2)" -> {"(AddEventHandler, 1)";}
"(Wrapf, 2)" -> {"(unmarshal, 3)";}
"(NewREST, 2)" -> {"(NewDefaultTableConvertor, 1)";}
"(Error, 0)" -> {"(Quote, 1)";"(format, 4)";}
"(Dgemm, 13)" -> {"(dgemmParallel, 12)";}
"(OpenContainer, 1)" -> {"(OpenComputeSystem, 2)";}
"(releaseEntry, 1)" -> {"(Put, 1)";"(Lock, 0)";"(log, 2)";"(Get, 0)";"(Unlock, 0)";"(Format, 4)";}
"(ref, 0)" -> {"(freeTrace, 1)";}
"(rebuild, 2)" -> {"(ReplaceOrInsert, 1)";}
"(Exp, 1)" -> {}
"(handleSignal, 1)" -> {"(DeliverSignal, 3)";}
"(PollUntil, 3)" -> {"(poller, 2)";}
"(newSession, 3)" -> {}
"(mkV2Events, 1)" -> {"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(Upload, 4)";"(sendConsumeCPURequests, 4)";}
"(Lvl, 1)" -> {}
"(isUnderlinedHeading, 1)" -> {"(RoundTripOpt, 2)";"(Dispatch, 4)";"(addEvent, 3)";"(roundTrip, 1)";}
"(int64, 1)" -> {}
"(keyIndex, 1)" -> {"(mkV2Events, 1)";}
"(RunOnce, 1)" -> {"(runOnce, 2)";}
"(GenerateAttachVolumeFunc, 2)" -> {}
"(Finish, 0)" -> {"(Nanoseconds, 0)";}
"(NewCustomResourceDefinitionHandler, 15)" -> {"(Store, 1)";}
"(ProtoReflect, 0)" -> {"(Resolve, 1)";"(SetUnknown, 1)";}
"(RawVector, 0)" -> {"(Untranspose, 0)";}
"(PathPrefix, 0)" -> {"(NewClientV3, 1)";}
"(NewChecker, 4)" -> {}
"(registerEvent, 3)" -> {"(addEventFile, 5)";}
"(NewUnsafeSet, 1)" -> {"(Init, 1)";}
"(ConfigureTransports, 1)" -> {"(configureTransports, 1)";}
"(dgemmParallel, 12)" -> {}
"(reset, 2)" -> {"(peekType, 0)";}
"(formatHeader, 3)" -> {"(Next, 0)";"(Write, 1)";}
"(Strings, 1)" -> {"(Reset, 0)";}
"(makeConnectedServerPipe, 0)" -> {}
"(clientHandshake, 2)" -> {"(clientAuthenticate, 1)";}
"(authCapabilityHandler, 1)" -> {}
"(Unlock, 0)" -> {"(writeback, 1)";}
"(Do, 2)" -> {"(Delims, 2)";}
"(addEvent, 3)" -> {}
"(GetNamespacesFromPodAffinityTerm, 2)" -> {"(getIfNewerThan, 2)";}
"(Clear, 0)" -> {"(isUnderlinedHeading, 1)";}
"(wrapperKeyBSONType, 1)" -> {"(dispatch, 0)";"(runSTM, 2)";"(NewClientConnectionWithPings, 2)";}
"(yaml_emitter_delete, 1)" -> {}
"(NewAggregator, 4)" -> {}
"(NewProxier, 13)" -> {}
"(AsFloat, 0)" -> {"(mustWrapValueReflect, 3)";}
"(failure, 1)" -> {"(ToJSONStream, 1)";}
"(callTTBFunc, 3)" -> {}
"(getClientConn, 3)" -> {"(getStartDialLocked, 1)";}
"(ReplaceOrInsert, 1)" -> {"(insert, 2)";}
"(DefineTests, 2)" -> {"(testScriptInPod, 4)";}
"(createProxier, 10)" -> {"(iptablesFlush, 1)";}
"(addServiceOnPortInternal, 4)" -> {"(ProxyLoop, 3)";}
"(monitorProbingStatus, 5)" -> {}
"(freepages, 0)" -> {}
"(sendConsumeCPURequests, 4)" -> {}
"(lockedUpdateLoadBalancerHosts, 2)" -> {"(GetLoadBalancer, 3)";}
"(unsafeCommit, 1)" -> {"(, 2)";}
"(Lock, 0)" -> {"(Ctx, 0)";}
"(newHTTP2Server, 2)" -> {}
"(AddWork, 3)" -> {"(getWrappedWorkerFunc, 1)";}
"(addMemberByURL, 3)" -> {"(listenGRPC, 0)";}
"(Launch, 0)" -> {"(NewClientHandler, 3)";}
"(AttachVolume, 2)" -> {"(GenerateAttachVolumeFunc, 2)";}
"(clientAuthenticate, 1)" -> {"(method, 0)";}
"(Start, 0)" -> {"(walkOperation, 1)";}
"(eof, 0)" -> {"(Detach, 2)";}
"(SetScale, 1)" -> {"(rescale, 1)";}
"(NewProxier, 6)" -> {"(createProxier, 6)";}
"(RuleProtocol, 1)" -> {"(RemoveInstance, 4)";}
"(deadlineLoop, 0)" -> {"(recvKeepAlive, 1)";}
"(handleAuth, 2)" -> {"(authCapabilityHandler, 1)";}
"(newAuthenticator, 2)" -> {"(initVerifier, 3)";}
"(AbsHostPath, 1)" -> {"(NewConnection, 2)";"(newConnection, 4)";}
"(createDialer, 0)" -> {"(ObserveDialLatency, 3)";}
"(waitForA, 1)" -> {"(failure, 1)";}
"(GetIntelRdtPath, 1)" -> {"(getIntelRdtRoot, 0)";}
"(DefaultControllerRateLimiter, 0)" -> {"(NewLimiter, 2)";}
"(applySnapshot, 2)" -> {}
"(generateEvent, 5)" -> {}
"(Unmarshal, 2)" -> {"(unmarshal, 3)";}
"(lookup, 0)" -> {"(compileUpdate, 1)";}
"(findJsonNameField, 1)" -> {"(allocate, 0)";}
"(Id, 1)" -> {"(topIndex, 0)";}
"(yaml_emitter_state_machine, 2)" -> {"(Gather, 0)";"(configureTransports, 1)";"(newWatcherGrpcStream, 1)";"(dispatch, 0)";"(StartRequest, 7)";"(resolveNow, 1)";"(NewClientConnectionWithPings, 2)";"(applyServiceConfigAndBalancer, 2)";}
"(IsUnknown, 1)" -> {}
"(doVolumeSetGetResponse, 2)" -> {"(DiscoverNode, 1)";}
"(GetTestImage, 1)" -> {}
"(NewClient, 2)" -> {"(defaultBaseTransport, 2)";}
"(HcsCloseProcess, 2)" -> {"(hcsCloseProcess, 1)";}
"(HasPrefix, 1)" -> {"(GetPrefix, 1)";}
"(createProxier, 6)" -> {"(platformDoesNotSupportError, 1)";}
"(convertNumber, 1)" -> {}
"(Snapshot, 2)" -> {}
"(monitorResizeEvents, 3)" -> {}
"(Errorf, 4)" -> {"(Error, 3)";}
"(Node, 0)" -> {"(newNode, 1)";}
"(CopyAllLogs, 4)" -> {}
"(NewTimeoutListener, 5)" -> {"(wrapTLS, 3)";}
"(AddInstances, 4)" -> {}
"(defaultBaseTransport, 2)" -> {"(appengineUrlfetchHook, 1)";}
"(writeLines, 3)" -> {"(CloseNotify, 0)";"(on100, 0)";"(SendSnapshot, 1)";"(mergeService, 1)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(newSnapshotReaderCloser, 2)";"(DeleteCollection, 4)";"(handleConnection, 2)";"(SyncServices, 1)";"(PortForward, 3)";"(HandlePodCleanups, 0)";"(startFrameWrite, 1)";"(NewClientConnectionWithPings, 2)";"(processPodOnNode, 5)";"(newDelayingQueue, 3)";"(removeAndReAdd, 1)";"(dispatch, 0)";"(AttachVolume, 2)";"(GenerateRegisterPluginFunc, 4)";}
"(format, 3)" -> {"(writeLines, 3)";}
"(WaitForAttachedDevice, 1)" -> {}
"(ProxyURL, 1)" -> {"(configureHTTP2Transport, 1)";}
"(Nanoseconds, 0)" -> {}
"(Equal, 4)" -> {"(formatUnequalValues, 2)";}
"(ToJSON, 1)" -> {"(SetBuffer, 1)";}
"(FindNamedType, 1)" -> {}
"(openDir, 1)" -> {}
"(tunnelHTTPConnect, 3)" -> {"(startHealthCheck, 1)";}
"(paragraph, 1)" -> {"(isUnderlinedHeading, 1)";}
"(Call, 5)" -> {"(RpcError_ErrorCode, 1)";"(GetDetail, 0)";}
"(updateEffectiveWindowSize, 0)" -> {"(lookup, 0)";}
"(Verify, 2)" -> {"(ecHash, 1)";}
"(Debug, 1)" -> {"(Properties, 4)";}
"(RemoveEndpoints, 3)" -> {"(doReconcile, 3)";}
"(newSnapshotSender, 4)" -> {}
"(Apiextensions, 0)" -> {"(New, 3)";}
"(LinesOf, 1)" -> {"(Line, 0)";}
"(NewDiscoveryController, 3)" -> {"(AddEventHandler, 1)";}
"(updateLastTransitionTime, 3)" -> {"(getIfNewerThan, 2)";}
"(timeoutFlush, 1)" -> {}
"(Logf, 3)" -> {"(releaseEntry, 1)";}
"(AddUint64, 2)" -> {"(AppendUint64, 1)";}
"(New, 6)" -> {"(newWatchableStore, 6)";}
"(canRetryError, 1)" -> {"(getStartDialLocked, 1)";}
"(initVerifier, 3)" -> {"(cancel, 2)";}
"(AfterSuiteDidRun, 1)" -> {"(handleSetupSummary, 2)";}
"(ecHash, 1)" -> {"(fromBase64, 1)";"(packKeyWire, 2)";}
"(Quote, 1)" -> {"(GRPCStatus, 0)";}
"(asyncFlush, 1)" -> {"(SearchFloat64s, 2)";}
"(mergeService, 1)" -> {"(addServicePortPortal, 5)";}
"(startDBSizeMonitorPerEndpoint, 2)" -> {}
"(NewVerifier, 3)" -> {}
"(doVolumeSet, 2)" -> {"(newWatcherGrpcStream, 1)";}
"(Reset, 0)" -> {"(Reset, 1)";"(Inf, 1)";"(MapRange, 0)";}
"(HTTPMembers, 0)" -> {"(isMembersEqual, 2)";}
"(get, 1)" -> {"(MapIndex, 1)";"(toMapKey, 1)";}
"(toMapKey, 1)" -> {"(retrieveUnexportedField, 3)";}
"(AddSpecific, 4)" -> {"(withRetries, 2)";}
"(getContainerStats, 1)" -> {"(Statistics, 0)";}
"(publicJWKSFromKeys, 1)" -> {"(Public, 0)";}
"(newFile, 2)" -> {"(loadPackage, 1)";}
"(NewV1PodResourcesServer, 3)" -> {"(serveUDP, 1)";}
"(walkPaths, 1)" -> {"(walkPathItem, 1)";}
"(enable, 0)" -> {}
"(uint64, 0)" -> {"(ReadUvarint, 1)";}
"(sendMergedSnap, 1)" -> {"(SendSnapshot, 1)";}
"(openExistingOrNew, 1)" -> {"(OpenFile, 3)";}
"(BeforeSuiteDidRun, 1)" -> {"(handleSetupSummary, 2)";}
"(acceptLoop, 0)" -> {}
"(NewEncoder, 1)" -> {"(New, 1)";"(commit, 1)";"(Write, 1)";"(newEncoderWithWriter, 1)";"(marshalDoc, 2)";"(setHttpHeader, 0)";"(NewEncoder, 1)";"(Caller, 1)";"(, 0)";"(Lock, 1)";"(Encode, 2)";"(Put, 1)";"(init, 0)";"(Unlock, 1)";"(Reset, 0)";"(Flush, 0)";"(marshal, 1)";}
"(removeAndReAdd, 1)" -> {}
"(Resolve, 1)" -> {"(FindNamedType, 1)";}
"(portForwardFunc, 3)" -> {"(WriteCloserWrapper, 1)";}
"(toRuleProtocol, 1)" -> {"(RuleProtocol, 1)";}
"(Offset, 1)" -> {"(quoteChar, 1)";}
"(newTimeBudget, 1)" -> {}
"(addPeerToProber, 6)" -> {"(monitorProbingStatus, 5)";}
"(KeysPrefix, 2)" -> {}
"(writeFrameFromHandler, 1)" -> {}
"(TailFile, 2)" -> {"(tailFileSync, 0)";}
"(StrategicMerge, 1)" -> {"(reload, 1)";}
"(newWatcherGrpcStream, 1)" -> {"(GetClientConn, 2)";"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";}
"(NewVersionedSet, 3)" -> {"(sortByLength, 2)";}
"(doReconcile, 3)" -> {"(Create, 2)";}
"(CancelRequest, 1)" -> {}
"(GetMetricWithLabelValues, 1)" -> {}
"(Install, 1)" -> {"(Consumes, 1)";}
"(Run, 4)" -> {"(runHTTPHandler, 3)";}
"(UpdateTransport, 4)" -> {"(updateTransport, 5)";}
"(panicMessage, 1)" -> {"(typeName, 0)";}
"(From, 0)" -> {"(Wrapf, 2)";}
"(newDecoratedWatcher, 2)" -> {}
"(StartCompactor, 3)" -> {}
"(Format, 4)" -> {"(Debug, 1)";}
"(splitPrefix, 1)" -> {"(newWatcherGrpcStream, 1)";}
"(Run, 3)" -> {"(waitForA, 1)";}
"(typeName, 0)" -> {}
"(ServeConn, 2)" -> {"(serve, 0)";}
"(AddNamedImport, 4)" -> {}
"(CreateVolume, 3)" -> {"(propsSlice, 1)";}
"(Connect, 0)" -> {"(connect, 0)";}
"(resetMessage, 1)" -> {"(SetUnknown, 1)";}
"(validate, 2)" -> {"(versionRange, 1)";}
"(newReadOnly, 1)" -> {"(LastIndex, 0)";"(maybeLastIndex, 0)";}
"(lockAndFlushAll, 0)" -> {"(writeDataFromHandler, 3)";"(newIPVSRequest, 1)";}
"(Panicf, 1)" -> {"(Clone, 0)";}
"(SyncServices, 1)" -> {}
"(InspectExec, 1)" -> {"(ContainerExecInspect, 2)";}
"(GetVMSetNames, 2)" -> {"(getAgentPoolScaleSets, 1)";}
"(stop, 0)" -> {"(DeleteProducer, 1)";}
"(ObserveDialLatency, 3)" -> {"(tunnelHTTPConnect, 3)";}
"(PutResources, 2)" -> {}
"(getJSONEncoder, 0)" -> {"(newWatcherGrpcStream, 1)";"(NewWithDelegate, 1)";"(roundTrip, 1)";"(NewDelayingQueueWithCustomClock, 2)";"(generateEvent, 5)";"(KeepAlive, 2)";"(SendSnapshot, 1)";"(applyServiceConfigAndBalancer, 2)";"(HandleStreams, 2)";"(newSnapshotReaderCloser, 2)";"(CloseNotify, 0)";"(handleConnection, 2)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(ServeConn, 2)";"(resolveNow, 1)";"(listenerRoutine, 0)";"(PortForward, 3)";"(recordToSink, 2)";"(makeConnectedServerPipe, 0)";"(on100, 0)";"(startFrameWrite, 1)";"(awaitOpenSlotForRequest, 1)";"(NewCacherFromConfig, 1)";"(newCluster, 2)";"(newCollector, 4)";"(apply, 3)";"(run, 3)";"(startStreamWriter, 6)";"(AddHTTP, 3)";"(DeleteCollection, 4)";"(NewServerTransport, 3)";"(scheduleBodyWrite, 0)";"(StartRequest, 7)";"(GetClientConn, 2)";}
"(ElectionServerToElectionClient, 1)" -> {"(stopNotify, 0)";}
"(VisitPodConfigmapNames, 2)" -> {"(visitContainerConfigmapNames, 2)";}
"(ProxyLoop, 3)" -> {"(getBackendConn, 7)";}
"(cancel, 0)" -> {"(Stop, 0)";}
"(Run, 2)" -> {"(runRaw, 2)";}
"(printNewLine, 0)" -> {"(tailFileSync, 0)";}
"(newClientConn, 2)" -> {}
"(UpdatePod, 1)" -> {"(syncPodFn, 1)";}
"(MessageOf, 1)" -> {"(Init, 1)";}
"(CompareAndSwapPointer, 3)" -> {}
"(peekType, 0)" -> {"(wrapperKeyBSONType, 1)";}
"(stopDBSizeMonitor, 0)" -> {"(UpdateEtcdDbSize, 2)";}
"(newBridge, 1)" -> {"(serveListen, 0)";}
"(addEventFile, 5)" -> {}
"(newLessor, 3)" -> {}
"(keyFn, 0)" -> {"(AddEventHandlerWithResyncPeriod, 2)";}
"(indirect, 1)" -> {"(eval, 1)";}
"(Statistics, 0)" -> {"(Properties, 2)";}
"(Terminate, 0)" -> {"(Terminate, 1)";}
"(Equal, 3)" -> {"(Equal, 4)";}
"(Delete, 1)" -> {"(Zero, 1)";}
"(reload, 1)" -> {"(recordToSink, 2)";"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(runOnce, 2)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(sendUpdates, 0)";"(NewProxier, 9)";"(ListenAndServe, 8)";"(StartInterceptingOutput, 0)";"(Run, 4)";"(GenerateBulkVolumeVerifyFunc, 4)";"(ListenAndServeReadOnly, 3)";}
"(verify, 2)" -> {"(keysFromRemote, 1)";}
"(newClient, 1)" -> {"(autoSync, 0)";}
"(New, 0)" -> {"(NewFileSet, 0)";}
"(WithStack, 1)" -> {}
"(Watch, 3)" -> {"(newWatcherGrpcStream, 1)";"(Connect, 0)";}
"(addServicePortPortal, 5)" -> {"(ProxyLoop, 3)";}
"(announceSetupFailure, 4)" -> {"(printFailure, 4)";}
"(WriteCloserWrapper, 1)" -> {"(ExecInContainer, 10)";}
"(newEncoderWithWriter, 1)" -> {"(yaml_emitter_set_unicode, 2)";}
"(BrowseDiagnosticLog, 3)" -> {"(newProperties, 1)";}
"(getMetricsFromNode, 2)" -> {}
"(iptablesFlush, 1)" -> {}
"(LeaseKeepAlive, 1)" -> {}
"(endBlock, 0)" -> {"(printDelimiter, 0)";}
"(fail, 2)" -> {"(NewEncoder, 1)";}
"(newProperties, 1)" -> {"(awaitOpenSlotForRequest, 1)";"(scheduleBodyWrite, 0)";"(GetClientConn, 2)";}
"(ServerPreferredResources, 1)" -> {"(fetchGroupVersionResources, 2)";}
"(getBackendConn, 7)" -> {}
"(walkResponse, 1)" -> {"(getJSONEncoder, 0)";}
"(awaitOpenSlotForRequest, 1)" -> {}
"(NewDesc, 4)" -> {"(labelPairSorter, 1)";}
"(closeRemoteChannels, 0)" -> {"(createCall, 5)";"(action, 1)";}
"(GC, 0)" -> {}
"(ForResource, 1)" -> {"(Samplecontroller, 0)";"(Flunders, 0)";}
"(To, 1)" -> {"(do, 3)";}
"(Delete, 2)" -> {"(processRaftRequest, 2)";}
"(doAppendInner, 2)" -> {"(isYesD, 0)";}
"(verifyDisksAttachedToNode, 2)" -> {}
"(escape, 1)" -> {"(printNewLine, 0)";}
"(Zero, 0)" -> {"(MessageOf, 1)";}
"(IsSelected, 1)" -> {"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(monitorResizeEvents, 3)";"(Launch, 0)";"(StartInterceptingOutput, 0)";"(NewProxier, 9)";"(ListenAndServeReadOnly, 3)";"(recordToSink, 2)";"(runOnce, 2)";"(sendUpdates, 0)";"(ListenAndServe, 8)";}
"(runHTTPHandler, 3)" -> {"(getHTTPRespBody, 1)";}
"(Handle, 2)" -> {"(Head, 2)";}
"(ListPodStatsAndUpdateCPUNanoCoreUsage, 0)" -> {"(listPodStats, 1)";}
"(getContainersToDeleteInPod, 3)" -> {"(updateStatusInternal, 3)";}
"(NewVersion, 1)" -> {"(FromGRPC, 1)";}
"(roundTrip, 1)" -> {"(handleReadLoopResponse, 1)";}
"(isLiteral, 1)" -> {"(cpuMask, 1)";}
"(NewStreamWatcher, 2)" -> {}
"(StartInterceptingOutput, 0)" -> {"(TailFile, 2)";}
"(StartRequest, 7)" -> {}
"(AnnounceSpecRunCompletion, 2)" -> {}
"(printFailure, 4)" -> {"(waitForChanges, 0)";}
"(processRaftRequest, 2)" -> {"(newWatcherGrpcStream, 1)";"(Connect, 0)";"(markDiskAsDeletable, 2)";"(parseProposeCtxErr, 2)";}
"(writeDataFromHandler, 3)" -> {"(writeFrameFromHandler, 1)";"(SortStrings, 1)";}
"(failureDetails, 1)" -> {"(endBlock, 0)";"(printSpecFailure, 4)";}
"(ioctlSetInt, 3)" -> {"(registerEvent, 3)";}
"(WithDeadline, 2)" -> {"(cancel, 2)";}
"(emitScalar, 8)" -> {"(callK, 1)";"(handleInterrupt, 0)";"(ConfigureTransports, 1)";}
"(SetError, 0)" -> {"(addEvent, 3)";}
"(SetBuffer, 1)" -> {"(constructBlock, 1)";}
"(ListenForSignal, 1)" -> {}
"(VolumesInUse, 2)" -> {"(volumesInUseFunc, 0)";}
"(unmarshal, 3)" -> {"(newParser, 1)";"(unmarshalPointer, 4)";}
"(Warn, 5)" -> {"(compact, 3)";}
"(isFeatureInRange, 2)" -> {}
"(errorS, 5)" -> {"(Insert, 1)";"(Remove, 1)";"(Do, 1)";"(Add, 1)";"(pathsToRemove, 2)";"(Add, 2)";"(Wait, 0)";"(Update, 3)";"(Reset, 0)";"(Flush, 0)";"(Transform, 3)";"(Send, 1)";"(openExistingOrNew, 1)";"(rotate, 0)";"(Header, 0)";"(append, 1)";"(Copy, 2)";"(New, 3)";"(Parse, 1)";"(Sync, 0)";"(Encode, 2)";"(Call, 5)";"(New, 0)";"(Decode, 1)";"(V, 1)";}
"(Set, 2)" -> {"(update, 4)";}
"(getStartDialLocked, 1)" -> {}
"(NormalizedName, 1)" -> {"(verify, 1)";}
"(fillClusterForMembers, 0)" -> {"(listenGRPC, 0)";}
"(msync, 1)" -> {"(newProxyServer, 3)";"(createCall, 5)";"(dialStreamLocal, 1)";"(monitorSize, 2)";"(Launch, 1)";"(createAndInitKubelet, 30)";"(SpecSuiteWillBegin, 2)";"(SpecSuiteDidEnd, 1)";"(hive, 1)";"(newConn, 2)";"(startKubelet, 6)";"(NewAggregator, 4)";"(RunOnce, 1)";}
"(NewEstablishingController, 2)" -> {"(DefaultControllerRateLimiter, 0)";}
"(fetchContainerStats, 1)" -> {"(Statistics, 0)";}
"(hcsCloseComputeSystem, 1)" -> {}
"(run, 3)" -> {"(StartGarbageCollection, 0)";"(sendUpdates, 0)";"(StartInterceptingOutput, 0)";"(CreateHealthCheck, 1)";"(NewTimeoutListener, 5)";"(ListenAndServeReadOnly, 3)";"(monitorResizeEvents, 3)";"(startFrameWrite, 1)";"(CloseNotify, 0)";"(AttachVolume, 2)";"(PortForward, 3)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(VerifySignature, 2)";"(Eventf, 7)";"(newSnapshotReaderCloser, 2)";"(runCmdContext, 2)";"(GenerateRegisterPluginFunc, 4)";"(createVolume, 2)";"(processPodOnNode, 5)";"(runOnce, 2)";"(Launch, 0)";"(on100, 0)";"(SendSnapshot, 1)";"(NewProxier, 9)";"(handleConnection, 2)";"(serve, 0)";"(Dispatch, 4)";"(HandlePodCleanups, 0)";"(verify, 1)";"(ListenAndServe, 8)";"(removeAndReAdd, 1)";"(SyncServices, 1)";"(lockedUpdateLoadBalancerHosts, 2)";"(generateEvent, 5)";"(NewMainKubelet, 30)";"(mergeService, 1)";"(createDialer, 0)";}
"(tearDown, 1)" -> {}
"(unlock, 0)" -> {"(ref, 0)";}
"(InitService, 1)" -> {}
"(handleCall, 1)" -> {"(decodeArguments, 3)";}
"(ListenAndServe, 0)" -> {"(serveUDP, 1)";}
"(newReflectCache, 0)" -> {"(Store, 1)";"(AsFloat, 0)";"(NewUnstructuredObjectTyper, 0)";}
"(NewSessionWithOptions, 1)" -> {"(newSession, 3)";}
"(resolveNow, 1)" -> {}
"(init, 0)" -> {"(AddExtensionType, 2)";}
"(iterate, 6)" -> {}
"(Update, 3)" -> {"(createMetadatas, 1)";}
"(processSignalResult, 2)" -> {}
"(cpuMask, 1)" -> {"(Make, 1)";}
"(truncatingFormat, 1)" -> {"(callTTBFunc, 3)";"(callTRFunc, 3)";}
"(MapRange, 0)" -> {"(closeRemoteChannels, 0)";}
"(NewWithDelegate, 1)" -> {}
"(New, 2)" -> {"(format, 2)";}
"(GracefulStop, 0)" -> {}
"(RoundTripOpt, 2)" -> {"(shouldRetryRequest, 3)";}
"(watchErrorStream, 2)" -> {}
"(mountServiceAccountToken, 2)" -> {"(createVolume, 2)";}
"(GetFullQualifiedPluginNameForVolume, 2)" -> {"(setNodeVolume, 4)";}
"(mustNewMember, 1)" -> {"(listenGRPC, 0)";}
"(newLoaderAtConfirmedDir, 4)" -> {"(HasPrefix, 1)";}
"(Create, 2)" -> {"(CreateUser, 3)";}
"(inBounds, 3)" -> {"(Run, 4)";"(runOnce, 2)";"(ListenAndServeReadOnly, 3)";"(handleConnection, 2)";"(cleanupOrphanedPodCgroups, 3)";"(GenerateRegisterPluginFunc, 4)";"(ListenAndServe, 8)";"(initForOS, 2)";"(addServiceOnPortInternal, 4)";"(NewMainKubelet, 30)";"(StartGarbageCollection, 0)";"(monitorResizeEvents, 3)";"(NewClientConnectionWithPings, 2)";"(createVolume, 2)";"(addServicePortPortal, 5)";"(sendUpdates, 0)";"(dispatch, 0)";"(Eventf, 7)";"(processPodOnNode, 5)";"(VerifySignature, 2)";"(NewClusterV3, 2)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(inBounds, 3)";"(RegisterPlugin, 3)";"(startDBSizeMonitorPerEndpoint, 2)";"(HandlePodCleanups, 0)";"(NewProxier, 9)";"(recordToSink, 2)";"(WaitForAttachedDevice, 1)";"(AddWork, 3)";"(startCompactorOnce, 2)";"(mergeService, 1)";"(UpdatePod, 1)";"(Launch, 0)";"(StartInterceptingOutput, 0)";"(Dispatch, 4)";"(verify, 1)";"(removeAndReAdd, 1)";"(SyncServices, 1)";"(Chan, 0)";"(AttachVolume, 2)";}
"(newSnapshotReaderCloser, 2)" -> {}
"(Execute, 2)" -> {"(Receive, 0)";}
"(rotate, 0)" -> {"(mill, 0)";}
"(After, 1)" -> {"(unlock, 0)";}
"(equalMessage, 2)" -> {"(equalUnknown, 2)";}
"(Header, 0)" -> {"(FromContext, 1)";"(finish, 1)";}
"(printf, 5)" -> {"(output, 6)";}
"(ServiceTicket, 3)" -> {"(SetThumbprint, 2)";}
"(labelPairSorter, 1)" -> {"(doAppendInner, 2)";}
"(newStore, 1)" -> {"(NewUnsafeSet, 1)";}
"(MigrateIfNeeded, 1)" -> {"(rollbackEtcd3MinorVersion, 2)";}
"(OperationCompleteHook, 2)" -> {"(setNodeVolume, 4)";}
"(Start, 1)" -> {"(walkPaths, 1)";}
"(Reset, 1)" -> {"(ProtoReflect, 0)";"(resetMessage, 1)";}
"(CreateUser, 3)" -> {"(roundTrip, 1)";"(startCompactorOnce, 2)";"(RoundTripOpt, 2)";"(UpdatePod, 1)";"(startDBSizeMonitorPerEndpoint, 2)";}
"(UpdateEtcdDbSize, 2)" -> {}
"(checkForUpdates, 1)" -> {"(getIfNewerThan, 2)";}
"(HcsCloseComputeSystem, 2)" -> {"(hcsCloseComputeSystem, 1)";}
"(Lock, 1)" -> {"(Ctx, 0)";"(waitDeletes, 4)";}
"(NewError, 3)" -> {"(validate, 2)";}
"(NewCRDFinalizer, 3)" -> {"(AddEventHandler, 1)";}
"(NewDiscoveryRESTMapper, 1)" -> {"(AddSpecific, 4)";}
"(Close, 0)" -> {"(Send, 1)";}
"(runRaw, 2)" -> {"(runWithFriendlyError, 3)";}
"(valueEncoder, 1)" -> {"(typeEncoder, 1)";}
"(addConnIfNeeded, 3)" -> {}
"(ExecInContainer, 10)" -> {"(InspectExec, 1)";}
"(Kill, 1)" -> {"(init, 0)";}
"(WaitForSingleObject, 2)" -> {}
"(Receive, 0)" -> {"(ParseNetlinkMessage, 1)";}
"(NewUnstructuredObjectTyper, 0)" -> {"(waitCancelSubstreams, 1)";"(getMetricsFromNode, 2)";"(newSession, 2)";"(DefineTests, 2)";"(SetUpAt, 2)";"(MigrateIfNeeded, 1)";"(newFilePipeline, 3)";"(newDecoratedWatcher, 2)";"(newConn, 2)";"(createMergedSnapshotMessage, 4)";"(PortForward, 4)";"(UpdateTransport, 4)";"(AddPod, 1)";"(BuildAuthn, 2)";"(receive, 0)";"(createCall, 5)";"(Evaluate, 1)";"(SpecSuiteDidEnd, 1)";"(StartTestServerOrDie, 4)";"(sendMergedSnap, 1)";"(startKubelet, 6)";"(CopyAllLogs, 4)";"(WatchPods, 4)";"(handleResponse, 2)";"(StartGatheringData, 0)";"(Dgemm, 13)";"(mountServiceAccountToken, 2)";"(passesVerifyFilter, 2)";"(copyStdout, 1)";"(RediscoverNode, 1)";"(copyStderr, 1)";"(watchErrorStream, 2)";"(Sink, 0)";"(handleResizes, 0)";"(createAndInitKubelet, 30)";"(freepages, 0)";"(NewStreamWatcher, 2)";"(Build, 3)";"(scheduleFrameWrite, 0)";"(DialWithConn, 4)";"(NewProxier, 13)";"(NewAggregator, 4)";"(initializeCSINode, 1)";"(waitForAvailableEndpoint, 1)";"(NewClientConn, 3)";"(newProxyServer, 3)";"(ListenForSignal, 1)";"(applyEntries, 2)";"(copyStdin, 0)";"(inWorker, 0)";"(monitorSize, 2)";"(Launch, 1)";"(Dispatch, 3)";"(RunOnce, 1)";"(SpecSuiteWillBegin, 2)";"(dialStreamLocal, 1)";"(createPD, 1)";"(recvLoop, 0)";"(podsRunning, 2)";"(newETCD3Storage, 2)";"(Verify, 2)";"(applySnapshot, 2)";"(sendLoop, 0)";}
"(eval, 1)" -> {"(getRequestedContainers, 2)";}
"(HandlePodCleanups, 0)" -> {"(GC, 0)";}
"(NewListener, 2)" -> {"(NewTimeoutListener, 5)";}
"(SyncPod, 4)" -> {"(containerStartSpec, 1)";}
"(log, 4)" -> {"(sweetenFields, 1)";}
"(fetch, 1)" -> {"(GetResponseRange, 0)";}
"(yaml_emitter_set_unicode, 2)" -> {"(TrimRightFunc, 2)";}
"(Get, 0)" -> {"(string, 1)";"(GetLatestPage, 0)";}
"(Errorf, 5)" -> {"(Warn, 5)";}
"(Sub, 1)" -> {"(AddUint64, 2)";}
"(newParser, 1)" -> {"(yaml_parser_set_input_string, 2)";}
"(recvKeepAlive, 1)" -> {"(sendKeepAliveLoop, 1)";}
"(newProxyServer, 3)" -> {"(NewProxier, 6)";}
"(NewFileSet, 0)" -> {"(Exec, 2)";"(OptionManager, 0)";}
"(mill, 0)" -> {}
"(startFrameWrite, 1)" -> {}
"(HandlePodReconcile, 1)" -> {"(deleteContainersInPod, 3)";}
"(markDiskAsDeletable, 2)" -> {"(Upload, 4)";}
"(OpenFile, 3)" -> {}
"(GetResponseRange, 0)" -> {"(checkTxnAuth, 3)";}
"(CanSet, 0)" -> {"(Evaluate, 1)";"(Dispatch, 3)";"(roundTrip, 1)";"(RoundTripOpt, 2)";"(newConn, 2)";"(dialStreamLocal, 1)";}
"(unregisterCallback, 1)" -> {"(HcsUnregisterComputeSystemCallback, 2)";}
"(kill, 0)" -> {"(Errorf, 5)";}
"(destroy, 0)" -> {"(yaml_emitter_delete, 1)";}
"(serve, 0)" -> {"(shutDownIn, 1)";}
"(ListenAndServe, 8)" -> {"(ListenAndServeKubeletServer, 10)";}
"(appengineUrlfetchHook, 1)" -> {"(FindDefaultCredentials, 2)";}
"(FindDefaultCredentials, 2)" -> {"(newSnapshotReaderCloser, 2)";"(CloseNotify, 0)";"(on100, 0)";"(startFrameWrite, 1)";"(PortForward, 3)";"(SendSnapshot, 1)";}
"(Detach, 2)" -> {"(doVolumeSet, 2)";}
"(addLivezChecks, 2)" -> {"(processPodOnNode, 5)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(handleConnection, 2)";"(SyncServices, 1)";"(HandlePodCleanups, 0)";"(mergeService, 1)";"(removeAndReAdd, 1)";"(AttachVolume, 2)";"(GenerateRegisterPluginFunc, 4)";}
"(Init, 1)" -> {"(CompareAndSwapPointer, 3)";}
"(waitForAvailableEndpoint, 1)" -> {}
"(receive, 0)" -> {}
"(handleInterrupt, 0)" -> {}
"(updateSecurityGroup, 4)" -> {"(createNodeSecurityGroup, 5)";}
"(popError, 1)" -> {"(WithLabelValues, 1)";}
"(HcsUnregisterComputeSystemCallback, 2)" -> {"(hcsUnregisterComputeSystemCallback, 1)";}
"(satisfiedExpectations, 0)" -> {"(poller, 2)";}
"(DeliverSignal, 3)" -> {"(deliver, 1)";}
"(RemoveContainer, 1)" -> {"(popError, 1)";}
"(cadvisorInfoToAcceleratorStats, 1)" -> {"(getRequestedContainers, 2)";"(OpenContainer, 1)";"(fetchContainerStats, 1)";"(getContainerStats, 1)";}
"(requestLogger, 2)" -> {"(handleAuth, 2)";}
"(wordMask, 1)" -> {"(effectivelyEquals, 1)";}
"(FileLine, 1)" -> {"(DialWithConn, 4)";}
"(maybeLastIndex, 0)" -> {"(RoundTripOpt, 2)";"(dispatch, 0)";"(runSTM, 2)";"(NewClientConnectionWithPings, 2)";"(roundTrip, 1)";}
"(DrainedAt, 0)" -> {"(RemoveImage, 1)";}
"(sortByLength, 2)" -> {}
"(ServerGroupsAndResources, 1)" -> {"(fetchGroupVersionResources, 2)";}
"(initForOS, 2)" -> {"(InitService, 1)";}
"(ChangeEvents, 2)" -> {}
"(do, 3)" -> {"(ErrAllFailed, 1)";}
"(SearchFloat64s, 2)" -> {"(Run, 4)";"(newWatcherGrpcStream, 1)";"(BulkVerifyVolumes, 1)";"(initForOS, 2)";}
"(QueryServiceConfig2, 5)" -> {"(clientConfig, 1)";}
"(newIPVSRequest, 1)" -> {"(writeLines, 3)";}
"(StartTestServer, 4)" -> {}
"(Gather, 0)" -> {}
"(StringAffineComparable, 1)" -> {"(doVolumeSetGetResponse, 2)";}
"(V1, 0)" -> {"(New, 3)";}
"(GetFileType, 1)" -> {"(getFileType, 1)";}
"(addCertRotation, 6)" -> {}
"(NewUnsecuredEtcd3TestClientServer, 1)" -> {"(PathPrefix, 0)";}
"(RegisterPlugin, 3)" -> {}
"(applyServiceConfigAndBalancer, 2)" -> {"(newCCBalancerWrapper, 3)";}
"(output, 1)" -> {"(objectType, 1)";}
"(sendConsumeCustomMetric, 5)" -> {}
"(syncPodFn, 1)" -> {"(wrapUp, 2)";}
"(ListenAndServePodResources, 4)" -> {"(NewV1PodResourcesServer, 3)";}
"(Now, 0)" -> {"(TimeFromUnixNano, 1)";}
"(applyEntries, 2)" -> {}
"(ListenAndServeReadOnly, 3)" -> {"(ListenAndServeKubeletReadOnlyServer, 5)";}
"(transformStartContainerError, 1)" -> {}
"(objectType, 1)" -> {"(dialStreamLocal, 1)";"(newConn, 2)";}
"(createPD, 1)" -> {"(CreatePD, 1)";}
"(CreatePD, 1)" -> {"(CreateDisk, 5)";}
"(selectLoadBalancer, 4)" -> {"(GetVMSetNames, 2)";}
"(GetPrefix, 1)" -> {"(Gather, 0)";"(Evaluate, 1)";"(StartTestServerOrDie, 4)";"(Dgemm, 13)";"(markDiskAsDeletable, 2)";"(mountServiceAccountToken, 2)";"(processPodOnNode, 5)";"(applySnapshot, 2)";"(handleResponse, 2)";"(watchErrorStream, 2)";"(createCall, 5)";"(newETCD3Storage, 2)";"(ConfigureTransports, 1)";"(Verify, 2)";"(newProxyServer, 3)";"(NewClientConn, 3)";"(NewListener, 2)";"(recvLoop, 0)";"(newConn, 2)";"(receive, 0)";"(addPeerToProber, 6)";"(ServeConn, 2)";"(createAndInitKubelet, 30)";"(newHTTP2Transport, 2)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(ListenForSignal, 1)";"(WatchPods, 4)";"(scheduleFrameWrite, 0)";"(Dispatch, 3)";"(ForResource, 1)";"(addMember, 1)";"(initializeCSINode, 1)";"(SetUpAt, 2)";"(run, 3)";"(NewAggregator, 4)";"(startPeer, 4)";"(createDialer, 0)";"(createPD, 1)";"(SpecSuiteDidEnd, 1)";"(mergeService, 1)";"(NewCacherFromConfig, 1)";"(resolveNow, 1)";"(HandlePodUpdates, 1)";"(CreateHealthCheck, 1)";"(startKubelet, 6)";"(removeAndReAdd, 1)";"(podsRunning, 2)";"(SpecSuiteWillBegin, 2)";"(RediscoverNode, 1)";"(RunOnce, 1)";"(GenerateRegisterPluginFunc, 4)";"(NewWithDelegate, 1)";"(handleConnection, 2)";"(monitorSize, 2)";"(Launch, 1)";"(handleResizes, 0)";"(getMetricsFromNode, 2)";"(callK, 1)";"(NewProxier, 13)";"(newDecoratedWatcher, 2)";"(NewClientConnectionWithPings, 2)";"(PortForward, 4)";"(copyStderr, 1)";"(AttachVolume, 2)";"(generateEvent, 5)";"(DeleteCollection, 4)";"(AddPod, 1)";"(Sink, 0)";"(newSession, 2)";"(BuildAuthn, 2)";"(sendLoop, 0)";"(Build, 3)";"(HandlePodReconcile, 1)";"(applyEntries, 2)";"(Run, 4)";"(newFilePipeline, 3)";"(freepages, 0)";"(copyStdin, 0)";"(copyStdout, 1)";"(passesVerifyFilter, 2)";"(applyServiceConfigAndBalancer, 2)";"(UpdateTransport, 4)";"(recordToSink, 2)";"(waitCancelSubstreams, 1)";"(dialStreamLocal, 1)";"(dispatch, 0)";"(NewDelayingQueueWithCustomClock, 2)";"(HandlePodCleanups, 0)";"(createMergedSnapshotMessage, 4)";"(MigrateIfNeeded, 1)";"(SyncServices, 1)";"(NewStreamWatcher, 2)";"(StartGatheringData, 0)";"(waitForAvailableEndpoint, 1)";"(CopyAllLogs, 4)";"(DefineTests, 2)";"(lockedUpdateLoadBalancerHosts, 2)";"(newWatcherGrpcStream, 1)";"(inWorker, 0)";"(listenerRoutine, 0)";"(sendMergedSnap, 1)";}
"(updateClientConnState, 1)" -> {"(applyServiceConfigAndBalancer, 2)";"(switchBalancer, 1)";"(newCCBalancerWrapper, 3)";}
"(checkTxnReqsPermission, 3)" -> {"(Connect, 0)";}
"(propsSlice, 1)" -> {}
"(Fatalf, 2)" -> {"(get, 0)";"(log, 4)";"(printf, 5)";"(FatalDepth, 2)";"(Logf, 3)";"(header, 2)";"(cancel, 0)";"(fail, 2)";"(Exit, 1)";}
"(NewTimer, 1)" -> {"(Now, 0)";}
"(getIntelRdtRoot, 0)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(updateTransport, 5)" -> {"(addCertRotation, 6)";}
"(close, 0)" -> {"(eof, 0)";}
"(newSession, 2)" -> {}
"(compileUpdate, 1)" -> {"(ParseServiceConfig, 1)";"(newWatcherGrpcStream, 1)";}
"(podsRunning, 2)" -> {}
"(sendUpdates, 0)" -> {}
"(setHttpHeader, 0)" -> {"(string, 1)";}
"(addMember, 1)" -> {"(HTTPMembers, 0)";"(addMemberByURL, 3)";}
"(Do, 3)" -> {"(debugResponse, 1)";}
"(newCallSet, 0)" -> {"(NewFileSet, 0)";}
"(waitDeletes, 4)" -> {"(waitDelete, 4)";}
"(GetUnknown, 0)" -> {"(createCall, 5)";"(freepages, 0)";}
"(inWorker, 0)" -> {"(handleCall, 1)";}
"(latestContainerStats, 1)" -> {"(UserDefinedMetricType, 1)";}
"(startCompactorOnce, 2)" -> {"(StartCompactor, 3)";}
"(Verifier, 1)" -> {"(NewVerifier, 3)";}
"(Marshal, 1)" -> {"(destroy, 0)";}
"(NewLimiter, 2)" -> {"(NewDelayingQueueWithCustomClock, 2)";}
"(toStringSlice, 1)" -> {"(QueryServiceConfig2, 5)";}
"(callK, 1)" -> {"(ClockWait, 3)";}
"(packKeyWire, 2)" -> {"(keysFromRemote, 1)";}
"(node, 2)" -> {"(emitScalar, 8)";}
"(format, 4)" -> {"(GRPCStatus, 0)";}
"(VerifySignature, 2)" -> {"(verify, 2)";}
"(SendMsg, 1)" -> {"(outPayload, 5)";}
"(GetFromPath, 1)" -> {"(NsHandle, 1)";}
"(Warningln, 2)" -> {"(newHTTP2Server, 2)";}
"(newHTTP2Client, 7)" -> {"(connectionErrorf, 5)";}
"(print, 4)" -> {"(LinesOf, 1)";"(Strings, 1)";"(String, 0)";"(runHandler, 1)";"(From, 1)";"(NodesOf, 1)";"(From, 0)";"(Infof, 2)";"(String, 2)";"(New, 1)";"(flushAll, 0)";"(add, 2)";"(Version, 0)";}
"(Lstat, 1)" -> {"(NewUnstructuredObjectTyper, 0)";}
"(byPrefixPriority, 1)" -> {}
"(newIntSet, 0)" -> {"(Max, 0)";}
"(sendSnap, 1)" -> {}
"(versionRange, 1)" -> {}
"(removeChild, 2)" -> {"(configureTransports, 1)";}
"(fromBase64, 1)" -> {"(keysFromRemote, 1)";}
"(hcsUnregisterComputeSystemCallback, 1)" -> {}
"(newETCD3Storage, 2)" -> {"(stopDBSizeMonitor, 0)";}
"(typeEncoder, 1)" -> {"(enable, 0)";}
"(newNode, 1)" -> {"(splitPrefix, 1)";}
"(finish, 1)" -> {"(SetError, 0)";}
"(ParseServiceConfig, 1)" -> {"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(sendConsumeCPURequests, 4)";}
"(NewServerTransport, 3)" -> {"(newHTTP2Server, 2)";}
"(setup, 2)" -> {"(ioctlSetInt, 3)";}
"(Run, 1)" -> {"(storeScratch, 4)";}
"(reset, 0)" -> {"(updateEffectiveWindowSize, 0)";}
"(deliver, 1)" -> {}
"(volumesInUseFunc, 0)" -> {}
"(getHTTPRespBody, 1)" -> {}
"(run, 0)" -> {"(processData, 0)";}
"(OptionManager, 0)" -> {"(newDelayingQueue, 3)";"(Example, 0)";"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(addListener, 1)" -> {"(newDelayingQueue, 3)";}
"(String, 2)" -> {"(grow, 2)";}
"(Chan, 0)" -> {}
"(walkOperation, 1)" -> {"(walkResponse, 1)";}
"(header, 2)" -> {"(formatHeader, 3)";}
"(NewCtxClient, 1)" -> {"(WithCancel, 1)";}
"(runCmdContext, 2)" -> {}
"(newCollector, 4)" -> {"(mapEventsToCustomEvents, 1)";}
"(getIfNewerThan, 2)" -> {}
"(InjectHandlers, 1)" -> {"(PushFrontNamed, 1)";}
"(setState, 1)" -> {}
"(DiscoverNode, 1)" -> {}
"(SortableMemberSliceByPeerURLs, 1)" -> {"(handleAuth, 2)";}
"(Get, 3)" -> {"(Do, 3)";}
"(retrieveUnexportedField, 3)" -> {"(UnsafeAddr, 0)";}
"(fdatasync, 1)" -> {"(msync, 1)";}
"(NewDelayingQueueWithCustomClock, 2)" -> {"(newDelayingQueue, 3)";}
"(SpecSuiteWillBegin, 2)" -> {"(escape, 1)";}
"(setNodeVolume, 4)" -> {"(verifyDisksAttachedToNode, 2)";}
"(printSpecFailure, 4)" -> {"(printFailure, 4)";}
"(keyToFilename, 1)" -> {"(Sum, 1)";}
"(Unknown, 1)" -> {"(IsUnknown, 1)";}
"(equalUnknown, 2)" -> {"(AttachVolume, 2)";"(handleConnection, 2)";"(NewDelayingQueueWithCustomClock, 2)";"(newWatcherGrpcStream, 1)";"(mergeService, 1)";"(removeAndReAdd, 1)";"(configureTransports, 1)";"(HandlePodCleanups, 0)";"(GenerateRegisterPluginFunc, 4)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(recordToSink, 2)";"(SyncServices, 1)";"(Run, 4)";"(processPodOnNode, 5)";}
"(configureTransports, 1)" -> {"(addConnIfNeeded, 3)";}
"(createMergedSnapshotMessage, 4)" -> {"(NewMessage, 3)";}
"(Public, 0)" -> {}
"(recordToSink, 2)" -> {}
"(scheduleFrameWrite, 0)" -> {"(isControl, 0)";}
"(handleSetupSummary, 2)" -> {"(failureDetails, 1)";"(announceSetupFailure, 4)";}
"(verify, 1)" -> {"(dedup, 1)";}
"(Retrieve, 0)" -> {"(profile, 0)";}
"(NewWithTLSConfig, 2)" -> {"(ProxyURL, 1)";}
"(mustWrapValueReflect, 3)" -> {"(runSTM, 2)";"(runOnce, 2)";"(newMux, 1)";"(StartTestServer, 4)";"(sendConsumeCPURequests, 4)";"(HandleStreams, 2)";"(newSnapshotReaderCloser, 2)";"(on100, 0)";"(rollbackEtcd3MinorVersion, 2)";"(NewClientConnectionWithPings, 2)";"(listenerRoutine, 0)";"(RoundTripOpt, 2)";"(newWatcherGrpcStream, 1)";"(monitorResizeEvents, 3)";"(SendSnapshot, 1)";"(NewProxier, 9)";"(Chan, 0)";"(CloseNotify, 0)";"(HandlePodCleanups, 0)";"(startFrameWrite, 1)";"(StartNode, 2)";"(DiscoverNode, 1)";"(NewDelayingQueueWithCustomClock, 2)";"(ListenAndServeReadOnly, 3)";"(Eventf, 7)";"(addPeerToProber, 6)";"(newHTTP2Transport, 2)";"(enable, 0)";"(NewMainKubelet, 30)";"(sendUpdates, 0)";"(sendConsumeMemRequests, 4)";"(processPodOnNode, 5)";"(VerifySignature, 2)";"(removeAndReAdd, 1)";"(newClientConn, 2)";"(DeleteCollection, 4)";"(StartGarbageCollection, 0)";"(startPeer, 4)";"(NewSession, 2)";"(sendConsumeCustomMetric, 5)";"(dgemmParallel, 12)";"(CreatePD, 1)";"(updateTransport, 5)";"(CreateVolume, 3)";"(handleConnection, 2)";"(startCompactorOnce, 2)";"(PortForward, 3)";"(UpdatePod, 1)";"(startDBSizeMonitorPerEndpoint, 2)";"(RestartNode, 1)";"(newLessor, 3)";"(mergeService, 1)";"(recordToSink, 2)";"(Run, 4)";"(apply, 3)";"(SyncServices, 1)";"(Launch, 0)";"(StartInterceptingOutput, 0)";"(Dispatch, 4)";"(keepAliveLoop, 2)";"(clientHandshake, 2)";"(verify, 1)";"(GetCollector, 1)";"(NewUnsecuredEtcd3TestClientServer, 1)";"(dispatch, 0)";"(ListenAndServe, 8)";"(AttachVolume, 2)";"(GenerateRegisterPluginFunc, 4)";"(handleSignal, 1)";"(roundTrip, 1)";"(createVolume, 2)";}
"(apply, 2)" -> {"(Panicf, 1)";}
"(NewReader, 4)" -> {"(newBpsLoop, 1)";}
"(TerminatePod, 1)" -> {"(updateStatusInternal, 3)";}
"(healthCheckFunc, 4)" -> {"(NewClientTransport, 7)";}
"(GetLoadBalancer, 3)" -> {"(getFloatingIPByPortID, 2)";}
"(add, 2)" -> {"(Add, 2)";}
"(WithCancel, 1)" -> {"(cancel, 2)";}
"(logName, 2)" -> {"(sendConsumeCPURequests, 4)";"(sendConsumeMemRequests, 4)";"(sendConsumeCustomMetric, 5)";"(CloseNotify, 0)";"(HandleStreams, 2)";"(runSTM, 2)";}
"(NewClusterAuthenticationTrustController, 2)" -> {"(keyFn, 0)";}
"(clientConfig, 1)" -> {"(connect, 0)";"(newETCD3HealthCheck, 1)";}
"(Store, 1)" -> {"(RetrieveOne, 4)";}
"(Bytes, 0)" -> {"(panicMessage, 1)";}
"(Make, 1)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(Consumes, 1)" -> {"(refreshMuxLocked, 0)";}
"(newMux, 1)" -> {"(newCond, 0)";}
"(NewMessage, 3)" -> {}
"(mapEventsToCustomEvents, 1)" -> {"(setup, 2)";}
"(Kill, 0)" -> {"(convertProcessError, 2)";}
"(openIDKeysetJSON, 1)" -> {"(publicJWKSFromKeys, 1)";}
"(dialStreamLocal, 1)" -> {}
"(createCall, 5)" -> {}
"(Samplecontroller, 0)" -> {}
"(newDecoder, 0)" -> {"(convertSliceNumbers, 2)";}
"(WithTTL, 1)" -> {"(mkSTM, 2)";}
"(PushFrontNamed, 1)" -> {}
"(LoadSources, 3)" -> {"(newFile, 2)";}
"(action, 1)" -> {}
"(method, 0)" -> {}
"(stopNotify, 0)" -> {}
"(Get, 2)" -> {"(reset, 2)";}
"(TrimRightFunc, 2)" -> {"(handleInterrupt, 0)";"(ConfigureTransports, 1)";"(callK, 1)";}
"(ContainerExecInspect, 2)" -> {}
"(newConn, 2)" -> {}
"(SendSnapshot, 1)" -> {"(sendSnap, 1)";}
"(Wait, 0)" -> {"(init, 0)";"(WaitForSingleObject, 2)";}
"(Parse, 1)" -> {"(ReduceFunc, 1)";}
"(Build, 2)" -> {}
"(enableCSM, 3)" -> {"(InjectHandlers, 1)";}
"(NodesOf, 1)" -> {"(Node, 0)";}
"(grow, 2)" -> {"(DecodeRuneInString, 1)";}
"(withRetries, 2)" -> {"(fetchGroupVersionResources, 2)";}
"(GetWorkerUnsafe, 1)" -> {"(getWrappedWorkerFunc, 1)";}
"(FromGRPC, 1)" -> {"(WithStack, 1)";}
"(runWithFriendlyError, 3)" -> {"(run, 3)";}
"(RawSyscallNoError, 4)" -> {"(RemoteCredProvider, 2)";}
"(tailFileSync, 0)" -> {"(waitForChanges, 0)";}
"(Initialize, 2)" -> {"(rebuild, 2)";}
"(InstallAPIGroups, 1)" -> {}
"(NewAddress, 1)" -> {"(updateResolverState, 2)";}
"(profile, 0)" -> {}
"(newETCD3HealthCheck, 1)" -> {}
"(Difference, 1)" -> {"(NewVersionedSet, 3)";}
"(WithLazyPrefix, 1)" -> {"(freepages, 0)";"(markDiskAsDeletable, 2)";}
"(connectionErrorf, 5)" -> {}
"(walkPathItem, 1)" -> {"(walkParameters, 1)";}
"(UnsafeAddr, 0)" -> {"(configureTransports, 1)";"(Upload, 4)";}
"(Zero, 1)" -> {"(CanSet, 0)";}
"(SetTypeForKey, 2)" -> {}
"(NewConditionController, 2)" -> {"(AddEventHandler, 1)";}
"(Stop, 0)" -> {"(New, 1)";"(Kill, 0)";"(Error, 0)";"(RemoveEndpoints, 3)";"(Do, 1)";"(Done, 0)";"(Close, 0)";"(kill, 0)";"(, 0)";"(stop, 0)";"(Kill, 1)";"(Signal, 1)";"(RemoveContainer, 1)";"(Wait, 0)";"(String, 0)";"(Unlock, 0)";"(NewTimer, 1)";"(Lock, 0)";"(Finish, 0)";"(After, 1)";"(Remove, 1)";"(Destroy, 0)";"(V, 1)";}
"(Iterate, 1)" -> {"(IterateUsing, 2)";}
"(runSTM, 2)" -> {}
"(Read, 1)" -> {"(Sum, 1)";}
"(rescale, 1)" -> {"(RediscoverNode, 1)";"(roundTrip, 1)";"(RoundTripOpt, 2)";"(Quo, 2)";"(newConn, 2)";"(dialStreamLocal, 1)";"(freepages, 0)";"(newWatcherGrpcStream, 1)";}
"(Max, 0)" -> {"(max, 1)";}
"(loadPackage, 1)" -> {"(NewChecker, 4)";}
"(remove, 1)" -> {"(wordMask, 1)";}
"(BuildAuthn, 2)" -> {}
"(AddEventHandlerWithResyncPeriod, 2)" -> {"(addListener, 1)";}
"(newConnection, 4)" -> {"(sendPings, 1)";}
"(getRequest, 1)" -> {"(newWatcherGrpcStream, 1)";"(Connect, 0)";}
"(KeepAlive, 2)" -> {"(deadlineLoop, 0)";}
"(WatchPods, 4)" -> {}
"(marshal, 1)" -> {"(reflectValue, 1)";}
"(NewClientConnectionWithPings, 2)" -> {"(newConnection, 4)";}
"(BulkVerifyVolumes, 1)" -> {"(setNodeVolume, 4)";}
"(syncPod, 1)" -> {"(SyncPod, 4)";}
"(Size, 0)" -> {"(GetLengthSid, 1)";}
"(refreshMuxLocked, 0)" -> {"(byPrefixPriority, 1)";}
"(Error, 3)" -> {"(errorS, 5)";}
"(singleRetrieve, 1)" -> {"(Retrieve, 0)";}
"(Dispatch, 4)" -> {}
"(RemoveInstance, 4)" -> {"(AddInstances, 4)";}
"(newWatchableStore, 6)" -> {"(syncVictimsLoop, 0)";}
"(ReadUvarint, 1)" -> {}
"(copyStdin, 0)" -> {}
"(Sum, 1)" -> {"(RawVector, 0)";}
"(SecondsSinceSSHKeySync, 0)" -> {"(addLivezChecks, 2)";}
"(PortForward, 3)" -> {"(portForwardFunc, 3)";}
"(wrapUp, 2)" -> {"(checkForUpdates, 1)";}
"(New, 1)" -> {"(newLoaderAtConfirmedDir, 4)";}
"(Terminate, 1)" -> {}
"(monitor, 0)" -> {}
"(serveConn, 1)" -> {}
"(RediscoverNode, 1)" -> {"(DiscoverNode, 1)";}
"(Flunders, 0)" -> {}
"(ReadAll, 1)" -> {"(readAll, 2)";}
"(constructBlock, 1)" -> {"(numLineBreaks, 1)";}
"(Ctx, 0)" -> {"(Delete, 2)";"(waitDeletes, 4)";}
"(updateStatusInternal, 3)" -> {"(updateLastTransitionTime, 3)";}
"(Put, 1)" -> {"(reset, 1)";"(reset, 2)";}
"(RemoteCredProvider, 2)" -> {"(executeCredentialProcess, 0)";"(createDialer, 0)";"(CreateHealthCheck, 1)";}
"(passesVerifyFilter, 2)" -> {"(verify, 1)";}
"(sync, 0)" -> {"(Observe, 1)";}
"(fn, 1)" -> {}
"(Write, 1)" -> {"(Send, 1)";"(printWithFileLine, 7)";}
"(Format, 1)" -> {"(ProtoReflect, 0)";}
"(ImportLayer, 4)" -> {"(layerPath, 2)";}
"(RestartNode, 1)" -> {}
"(evictOldest, 1)" -> {"(newWatcherGrpcStream, 1)";}
"(DeleteCollection, 4)" -> {}
"(NewClientTransport, 7)" -> {"(newHTTP2Client, 7)";}
"(ReduceFunc, 1)" -> {"(Offset, 1)";"(Id, 1)";}
"(TimeFromUnixNano, 1)" -> {}
"(isYesD, 0)" -> {"(NewClientConnectionWithPings, 2)";"(dispatch, 0)";}
"(newCollector, 2)" -> {"(mapEventsToCustomEvents, 1)";}
"(createNodeSecurityGroup, 5)" -> {"(toRuleProtocol, 1)";}
"(commit, 1)" -> {"(unsafeCommit, 1)";}
"(evict, 0)" -> {"(evictOldest, 1)";}
"(Destroy, 0)" -> {"(GetPath, 0)";}
"(timeoutIO, 0)" -> {"(configureTransports, 1)";"(enable, 0)";"(RoundTripOpt, 2)";"(StartRequest, 7)";"(resolveNow, 1)";"(NewClientConnectionWithPings, 2)";"(roundTrip, 1)";"(applyServiceConfigAndBalancer, 2)";"(Gather, 0)";"(newWatcherGrpcStream, 1)";"(dispatch, 0)";}
"(getDelegate, 0)" -> {"(NewDiscoveryRESTMapper, 1)";}
"(HandlePodUpdates, 1)" -> {"(dispatchWork, 4)";}
"(Delims, 2)" -> {"(roundTrip, 1)";"(RoundTripOpt, 2)";}
"(DecodeRuneInString, 1)" -> {"(errorf, 3)";"(GetUnknown, 0)";"(SetError, 0)";}
"(getTrimmed, 1)" -> {"(timeoutIO, 0)";}
"(quoteChar, 1)" -> {"(Dispatch, 4)";}
"(keysFromRemote, 1)" -> {}
"(Send, 1)" -> {"(SendMsg, 1)";}
"(Inf, 1)" -> {"(closeRemoteChannels, 0)";}
"(Revoke, 2)" -> {"(Values, 0)";}
"(pingTimeoutSeconds, 0)" -> {"(configureTransports, 1)";}
"(dispatchWork, 4)" -> {"(TerminatePod, 1)";}
"(copyStdout, 1)" -> {}
"(unmarshalPointer, 4)" -> {"(initializeCSINode, 1)";}
"(GetDetail, 0)" -> {"(CancelRequest, 1)";}
"(NewClientConn, 3)" -> {"(newMux, 1)";}
"(SetUpAt, 2)" -> {"(totalSecretBytes, 1)";}
"(marshalDoc, 2)" -> {"(nodev, 1)";}
"(NewLeaseFromLeaseClient, 3)" -> {"(WithCancel, 1)";}
"(GetWithContext, 1)" -> {"(singleRetrieve, 1)";}
"(SpecDidComplete, 1)" -> {"(failureDetails, 1)";}
"(ListenAndServeKubeletReadOnlyServer, 5)" -> {"(ListenAndServe, 0)";}
"(create, 3)" -> {"(logName, 2)";}
"(keepAliveLoop, 2)" -> {"(replyToClient, 2)";}
"(parse, 1)" -> {"(NewKey, 2)";}
"(WithLabelValues, 1)" -> {"(GetMetricWithLabelValues, 1)";}
"(log, 2)" -> {"(write, 0)";}
"(Build, 3)" -> {}
"(cadvisorInfoToNetworkStats, 1)" -> {"(latestContainerStats, 1)";}
"(newHTTP2Transport, 2)" -> {"(Warningln, 2)";}
"(Exit, 1)" -> {"(print, 4)";}
"(MapIndex, 1)" -> {"(equalMessage, 2)";}
"(AppendUint, 1)" -> {"(dispatch, 0)";"(NewClientConnectionWithPings, 2)";}
"(ToFloat64, 1)" -> {"(ToFloat64E, 1)";}
"(flushAll, 0)" -> {"(Sync, 0)";}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 0a839c6c3b9489716e216b4fc9dd2f9c17948c16
| test | kubernetes kubernetes pkg controller endpointslice reconciler test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to slice at line may start a goroutine click here to show the line s of go which triggered the analyzer go for slice range slices if strings hasprefix slice name testcase service name t fatalf expected endpointslice name to start with s got s testcase service name slice name if reflect deepequal testcase expectedlabels slice labels t errorf expected endpointslice to have labels v got v testcase expectedlabels slice labels if slice labels testcase service name t fatalf expected endpointslice to have label set with s value got s testcase service name slice labels if slice annotations triggertime format time t fatalf expected endpointslice trigger time annotation to be s got s triggertime format time slice annotations validate that this slice has address type matching expected expectedendpointlist testcase expectedendpointperslice if expectedendpointlist nil t fatalf address type v is not expected slice addresstype if len slice endpoints len expectedendpointlist t fatalf expected v endpoint got d len expectedendpointlist len slice endpoints test is limited to one endpoint endpoint slice endpoints if reflect deepequal endpoint expectedendpointlist t fatalf expected endpoint v got v expectedendpointlist endpoint expecttrackedresourceversion t r endpointslicetracker slice expectmetrics t expectedmetrics desiredslices actualslices desiredendpoints addedpersync len testcase expectedendpointperslice removedpersync numcreated len testcase expectedendpointperslice numupdated numdeleted click here to show extra information the analyzer produced the following dot graph describes paths through the callgraph that could lead to a function calling a goroutine digraph g writeback merge schedulebodywrite shutdownin sendpings withtimeout withdeadline example rpcerror errorcode cancelrequest retrieveone sendconsumecustommetric serverpreferredresources rediscovernode run sendloop setupat newconn createmergedsnapshotmessage createandinitkubelet newclientconn evaluate specdidcomplete listenandservereadonly restartnode mountserviceaccounttoken newsession handlepodreconcile deletecollection discovernode sendconsumememrequests configuretransports beforesuitedidrun keepaliveloop specsuitewillbegin specsuitedidend apply dispatch handleresponse recvloop build processpodonnode sink newdelayingqueuewithcustomclock roundtripopt closenotify startcompactoronce getcollector chan addpeertoprober newclientconnectionwithpings portforward createvolume newproxier newstreamwatcher eventf listenerroutine verifysignature updatepod newmainkubelet sendmergedsnap receive starttestserver dispatch handlestreams mergeservice dispatch runstm startgarbagecollection sendupdates roundtrip updatetransport monitorresizeevents launch updatetransport markdiskasdeletable applysnapshot enable recordtosink newlessor handlepodcleanups runonce portforward getmetricsfromnode newaggregator newproxier newclientconn runonce attachvolume dialstreamlocal addpod newsnapshotreadercloser sendconsumecpurequests aftersuitedidrun newfilepipeline startframewrite buildauthn newmux dgemmparallel sendsnapshot monitorsize podsrunning generateregisterpluginfunc applyentries newwatchergrpcstream handlepodupdates createpd verify newdecoratedwatcher handleconnection launch scheduleframewrite handlesignal removeandreadd callk newproxyserver listenandserve syncservices startinterceptingoutput createcall createvolume verify startkubelet startnode startpeer waitcancelsubstreams initializecsinode startdbsizemonitorperendpoint clienthandshake updateobjectcount newtimebudget addhttp search execute fromcontext unknown calltrfunc replytoclient deadlineloop runonce getfailedcontainers startcaching convertprocesserror processsignalresult processdata oneachwrite syncvictimsloop movevictims cleanuporphanedpodcgroups installapis installapigroups newticker yaml emitter state machine togrpc pathstoremove difference lastindex passesverifyfilter initializecsinode dispatch runstm freepages newclientconnectionwithpings roundtrip normalizedname addpod roundtripopt tokenvolumesource newaffinityterm getnamespacesfrompodaffinityterm waitdelete watch freepages wraperr newmainkubelet defaultnodestatusfuncs serveudp spawnworker string decoderuneinstring encode msgheader legacylogsymlink logsymlink getagentpoolscalesets createorupdatevmss msgheader do wraperr handleconnection exec abshostpath removeimage removeimage standardmethodargumentdecode logsymlink transformstartcontainererror createcomputesystem specsuitedidend announcespecruncompletion from to tojsonstream chan verifysignature serve lockedupdateloadbalancerhosts startframewrite closenotify createvolume newtimeoutlistener eventf newsnapshotreadercloser portforward sendsnapshot dispatch ornil add evict version newversion getrequest createhealthcheck healthchecks togrpc electionservertoelectionclient grpcstatus exp settypeforkey getfiletype erorr processsignalresult startpeer newsnapshotsender defaultnodestatusfuncs volumesinuse newproxier newcustomproxier layerpath newopenidmetadata openidkeysetjson sync errnoerr newcluster fillclusterformembers iterateusing eachmapentry removeblock clear paragraph clone newtls evaluate waitforchanges changeevents sign run head keyindex pingtimeoutseconds config tostringslice new processflag hcscloseprocess sortstrings topindex dispatch max updateresolverstate updateclientconnstate numlinebreaks newwatchergrpcstream servelisten serveconn monitorsize monitorresizeevents empty stringaffinecomparable dispatch eventf handlereadloopresponse append flushcoldbuf mayberotatestreams sendloop requestprogress opencomputesystem getclientconn getclientconn tunnelsynchealthchecker secondssincesshkeysync done erorr closenotify switchbalancer newccbalancerwrapper getfailedcontainers syncpod processflag setstate errallfailed verify startgatheringdata removeimage imageremove zcopy browsediagnosticlog hive startgarbagecollection sendupdates newproxier listenandserve listenandservereadonly newmainkubelet runonce monitorresizeevents launch startinterceptingoutput cancel removechild newcacherfromconfig startcaching keys keysprefix freetrace provider resettransport starthealthcheck nodev node launch waitversion walkparameters handlestreams startrequest resolvenow newservertransport applyserviceconfigandbalancer makeconnectedserverpipe defaultcollector newconn dialstreamlocal createcall startstreamwriter ornil getrequestedcontainersinfo ornil effectivelyequals isselected pollimmediateuntil polluntil remove removeblock errorf createhealthcheck executecredentialprocess createcall configuretransports rediscovernode createdialer executecredentialprocess reflectvalue valueencoder getlatestpage rawsyscallnoerror sendrequest fn get getusing createandinitkubelet startgarbagecollection newtls addhttp mustnewmember launch startstreamwriter deletecontainersinpod getcontainerstodeleteinpod newclienthandler requestlogger createorupdatevmss putresources update canset yaml parser set input string debugresponse newteereader mkstm runstm awaitopenslotforrequest schedulebodywrite getclientconn write fdatasync newfilepipeline newccbalancerwrapper build handler lstat formatunequalvalues truncatingformat v setv getlengthsid getfiletype convertslicenumbers convertnumber runhandler handler do readall parseproposectxerr upload generatebulkvolumeverifyfunc operationcompletehook getfullqualifiedpluginnameforvolume shutdown copystderr getfloatingipbyportid selectloadbalancer set isliteral checktxnauth checktxnreqspermission starttestserverordie starttestserver setendpoints newaddress kind ok satisfiedexpectations recvloop iscontrol add setscale signal marshal deleteproducer next lookup nshandle healthchecks listenerroutine makeconnectedserverpipe do printdepth filter output bytes getbuffer sendconsumememrequests getrequestedcontainers executeprestophook properties defaultcollector readall setunknown autosync lvl get newerror insert descend addpod visitpodconfigmapnames newbpsloop newbackend run addextensiontype gettrimmed newticker yaml emitter emit unlock delete getwrappedworkerfunc newkey createmetadatas withttl appenduint apply addnamedimport getcgroupstats cadvisorinfotonetworkstats snapshot newtlslistener acceptloop transform kind mayberotatestreams descend newconnection monitor previousdestroy updateobjectcount waitcancelsubstreams listpodstats makepodstoragestats output timeoutflush format reset error dispatch dispatch provider newwatchergrpcstream dispatch listpodstats listpodstats dialwithconn newmux handlestreams attachvolume chan removeandreadd deletecollection startframewrite dispatch newmainkubelet sendupdates handlepodcleanups generateregisterpluginfunc runstm syncservices startinterceptingoutput starttestserver closenotify sendconsumecustommetric createpd startdbsizemonitorperendpoint launch mergeservice dispatch sendconsumememrequests newproxier handleconnection createvolume newsnapshotreadercloser dgemmparallel updatetransport newdelayingqueuewithcustomclock listenandserve sendconsumecpurequests listenandservereadonly eventf portforward createvolume verifysignature newclientconnectionwithpings startcompactoronce sendsnapshot processpodonnode newwatchergrpcstream recordtosink run updatepod monitorresizeevents discovernode clienthandshake startgarbagecollection runonce handlesignal verify fetchgroupversionresources eachmapentry dialstreamlocal launch build createandinitkubelet updatetransport newfilepipeline newsession buildauthn startnode starttestserverordie callk migrateifneeded inworker sink receive addpeertoprober mountserviceaccounttoken newconn newproxier setupat rediscovernode verify copystdin startgatheringdata enable servergroupsandresources newaggregator waitforavailableendpoint newclientconn roundtrip handleresizes evaluate sendloop roundtripopt dispatch createcall freepages copystderr initializecsinode copyalllogs newclientconn dgemm applysnapshot copystdout createmergedsnapshotmessage scheduleframewrite passesverifyfilter listenforsignal recvloop newsession runonce listenerroutine newproxyserver watcherrorstream sendmergedsnap podsrunning handleresponse specsuitewillbegin specsuitedidend portforward waitcancelsubstreams applyentries startkubelet startpeer createpd restartnode newdecoratedwatcher addpod getmetricsfromnode watchpods getcollector newstreamwatcher monitorsize definetests newlessor createcomputesystem values retrieveunexportedfield yaml emitter emit handleinterrupt configuretransports callk newteereader newwatchergrpcstream awaitopenslotforrequest schedulebodywrite connect getclientconn flush split lockandflushall split writedatafromhandler portforward portforward insert newnamingconditioncontroller addeventhandler sweetenfields append remove pathstoremove wait openexistingornew header decode do v transform sync call flush index flushcoldbuf new add parse new update add send put copy rotate encode reset newcontroller newcallset client newclient dedup loadsources infof printf sendkeepaliveloop leasekeepalive oneachwrite sendsnapshot newsnapshotreadercloser closenotify startframewrite portforward delete empty testscriptinpod gettestimage getusing findjsonnamefield outpayload finishstream sendconsumecpurequests closenotify deletecollection sendconsumecustommetric runstm generatebulkvolumeverifyfunc newclientconnectionwithpings run newwatchergrpcstream recordtosink addevent newdelayingqueuewithcustomclock roundtripopt dispatch sendconsumememrequests handlestreams roundtrip markdiskasdeletable quo getclientconn dispatch awaitopenslotforrequest schedulebodywrite newclientconnectionwithpings startnode newsession keepalive createvolume tokenvolumesource load start get next pollimmediateuntil newdiscoverycontroller stopservers newbackend sync encode newrest commit add newwithtlsconfig read setendpoints node getdelegate listpodstatsandupdatecpunanocoreusage from keytofilename start unmarshal hcscloseprocess set handle new terminate verifier get write importlayer search remove withtimeout listpodstats newnamingconditioncontroller new initialize serviceticket ok newcustomresourcedefinitionhandler get sendrequest close lock newcrdfinalizer shutdown newestablishingcontroller reset config newclient get get delete sub v enablecsm newdesc installapis fetch reset zero newconditioncontroller teardown wait decode getfrompath do keys unregistercallback newclient add do remove apiextensions polluntil run new delete run run apply add newstore transform install new parse newclusterauthenticationtrustcontroller new newsessionwithoptions sign equal sync flush newopenidmetadata getcgroupstats reset get client parse put newctxclient size init hcsclosecomputesystem send newkubernetesapiapprovalpolicyconformantconditioncontroller format newauthenticator getwithcontext unlock revoke newleasefromleaseclient error tojson installtunneler newcontroller poller awaitrequestcancel newcustomproxier createproxier clockwait newdefaulttableconvertor previousdestroy index insert untranspose dispatch newclientconnectionwithpings closenotify handlestreams newwatchergrpcstream deletecollection sendconsumecpurequests sendconsumememrequests sendconsumecustommetric userdefinedmetrictype getrequestedcontainers garbagecollect drainedat listenandservekubeletserver listenandserve listenandservepodresources listenandservepodresources printwithfileline create allocate roundtrip getrequestedcontainersinfo fetchcontainerstats roundtripopt callttbfunc apply totalsecretbytes generateregisterpluginfunc registerplugin containerstartspec legacylogsymlink printdelimiter changeevents updateloadbalancer updatesecuritygroup copy zcopy startgarbagecollection garbagecollect spawnworker stopservers gracefulstop string iterate makepodstoragestats cadvisorinfotoacceleratorstats newdelayingqueue processpodonnode getworkerunsafe platformdoesnotsupporterror isfeatureinrange fataldepth printdepth handleresponse awaitrequestcancel getbuffer reset initializecsinode wraptls newtlslistener expecttrackedresourceversion fatalf errorf merge withlazyprefix handleresizes requestprogress properties errnoerr strategicmerge msync getfiletype movevictims newdelayingqueue updateloadbalancer getloadbalancer setv fileline compact launch newproxier listenandserve recordtosink newmainkubelet startgarbagecollection runonce monitorresizeevents sendupdates startinterceptingoutput listenandservereadonly setunknown newclientconnectionwithpings markdiskasdeletable dispatch opendir newcond method format format caller next parsenetlinkmessage printf output ismembersequal sortablememberslicebypeerurls getcollector newcollector visitcontainerconfigmapnames newaffinityterm newintset installtunneler tunnelsynchealthchecker storescratch inbounds add ref indirect descend iterate sink shouldretryrequest canretryerror filter set setthumbprint addeventhandler addeventhandlerwithresyncperiod connect resettransport createdisk listengrpc newbridge imageremove executeprestophook get load getpath getintelrdtpath reset newreadonly handlestreams waitversion ismembersequal starthealthcheck healthcheckfunc decode newdecoder observe asyncflush startkubelet listenandservepodresources upload newreader decodearguments standardmethodargumentdecode newreflectcache line lookup finishstream addevent newkubernetesapiapprovalpolicyconformantconditioncontroller addeventhandler wrapf unmarshal newrest newdefaulttableconvertor error quote format dgemm dgemmparallel opencontainer opencomputesystem releaseentry put lock log get unlock format ref freetrace rebuild replaceorinsert exp handlesignal deliversignal polluntil poller newsession sendconsumememrequests sendconsumecustommetric closenotify handlestreams upload sendconsumecpurequests lvl isunderlinedheading roundtripopt dispatch addevent roundtrip keyindex runonce runonce generateattachvolumefunc finish nanoseconds newcustomresourcedefinitionhandler store protoreflect resolve setunknown rawvector untranspose pathprefix newchecker registerevent addeventfile newunsafeset init configuretransports configuretransports dgemmparallel reset peektype formatheader next write strings reset makeconnectedserverpipe clienthandshake clientauthenticate authcapabilityhandler unlock writeback do delims addevent getnamespacesfrompodaffinityterm getifnewerthan clear isunderlinedheading wrapperkeybsontype dispatch runstm newclientconnectionwithpings yaml emitter delete newaggregator newproxier asfloat mustwrapvaluereflect failure tojsonstream callttbfunc getclientconn getstartdiallocked replaceorinsert insert definetests testscriptinpod createproxier iptablesflush addserviceonportinternal proxyloop monitorprobingstatus freepages sendconsumecpurequests lockedupdateloadbalancerhosts getloadbalancer unsafecommit lock ctx addwork getwrappedworkerfunc addmemberbyurl listengrpc launch newclienthandler attachvolume generateattachvolumefunc clientauthenticate method start walkoperation eof detach setscale rescale newproxier createproxier ruleprotocol removeinstance deadlineloop recvkeepalive handleauth authcapabilityhandler newauthenticator initverifier abshostpath newconnection newconnection createdialer observediallatency waitfora failure getintelrdtpath getintelrdtroot defaultcontrollerratelimiter newlimiter applysnapshot generateevent unmarshal unmarshal lookup compileupdate findjsonnamefield allocate id topindex yaml emitter state machine gather configuretransports newwatchergrpcstream dispatch startrequest resolvenow newclientconnectionwithpings applyserviceconfigandbalancer isunknown dovolumesetgetresponse discovernode gettestimage newclient defaultbasetransport hcscloseprocess hcscloseprocess hasprefix getprefix createproxier platformdoesnotsupporterror convertnumber snapshot monitorresizeevents errorf error node newnode copyalllogs newtimeoutlistener wraptls addinstances defaultbasetransport appengineurlfetchhook writelines closenotify sendsnapshot mergeservice newsnapshotreadercloser deletecollection handleconnection syncservices portforward handlepodcleanups startframewrite newclientconnectionwithpings processpodonnode newdelayingqueue removeandreadd dispatch attachvolume generateregisterpluginfunc format writelines waitforattacheddevice proxyurl nanoseconds equal formatunequalvalues tojson setbuffer findnamedtype opendir tunnelhttpconnect starthealthcheck paragraph isunderlinedheading call rpcerror errorcode getdetail updateeffectivewindowsize lookup verify echash debug properties removeendpoints doreconcile newsnapshotsender apiextensions new linesof line newdiscoverycontroller addeventhandler updatelasttransitiontime getifnewerthan timeoutflush logf releaseentry new newwatchablestore canretryerror getstartdiallocked initverifier cancel aftersuitedidrun handlesetupsummary echash packkeywire quote grpcstatus asyncflush mergeservice addserviceportportal startdbsizemonitorperendpoint newverifier dovolumeset newwatchergrpcstream reset reset inf maprange httpmembers ismembersequal get mapindex tomapkey tomapkey retrieveunexportedfield addspecific withretries getcontainerstats statistics publicjwksfromkeys public newfile loadpackage serveudp walkpaths walkpathitem enable readuvarint sendmergedsnap sendsnapshot openexistingornew openfile beforesuitedidrun handlesetupsummary acceptloop newencoder new commit write newencoderwithwriter marshaldoc sethttpheader newencoder caller lock encode put init unlock reset flush marshal removeandreadd resolve findnamedtype portforwardfunc writecloserwrapper toruleprotocol ruleprotocol offset quotechar newtimebudget addpeertoprober monitorprobingstatus keysprefix writeframefromhandler tailfile tailfilesync strategicmerge reload newwatchergrpcstream getclientconn awaitopenslotforrequest schedulebodywrite newversionedset sortbylength doreconcile create cancelrequest getmetricwithlabelvalues install consumes run runhttphandler updatetransport updatetransport panicmessage typename from wrapf newdecoratedwatcher startcompactor format debug splitprefix newwatchergrpcstream run waitfora typename serveconn serve addnamedimport createvolume propsslice connect connect resetmessage setunknown validate versionrange newreadonly lastindex maybelastindex lockandflushall writedatafromhandler newipvsrequest panicf clone syncservices inspectexec containerexecinspect getvmsetnames getagentpoolscalesets stop deleteproducer observediallatency tunnelhttpconnect putresources getjsonencoder newwatchergrpcstream newwithdelegate roundtrip newdelayingqueuewithcustomclock generateevent keepalive sendsnapshot applyserviceconfigandbalancer handlestreams newsnapshotreadercloser closenotify handleconnection serveconn resolvenow listenerroutine portforward recordtosink makeconnectedserverpipe startframewrite awaitopenslotforrequest newcacherfromconfig newcluster newcollector apply run startstreamwriter addhttp deletecollection newservertransport schedulebodywrite startrequest getclientconn electionservertoelectionclient stopnotify visitpodconfigmapnames visitcontainerconfigmapnames proxyloop getbackendconn cancel stop run runraw printnewline tailfilesync newclientconn updatepod syncpodfn messageof init compareandswappointer peektype wrapperkeybsontype stopdbsizemonitor updateetcddbsize newbridge servelisten addeventfile newlessor keyfn addeventhandlerwithresyncperiod indirect eval statistics properties terminate terminate equal equal delete zero reload recordtosink newmainkubelet startgarbagecollection runonce monitorresizeevents launch sendupdates newproxier listenandserve startinterceptingoutput run generatebulkvolumeverifyfunc listenandservereadonly verify keysfromremote newclient autosync new newfileset withstack watch newwatchergrpcstream connect addserviceportportal proxyloop announcesetupfailure printfailure writecloserwrapper execincontainer newencoderwithwriter yaml emitter set unicode browsediagnosticlog newproperties getmetricsfromnode iptablesflush leasekeepalive endblock printdelimiter fail newencoder newproperties awaitopenslotforrequest schedulebodywrite getclientconn serverpreferredresources fetchgroupversionresources getbackendconn walkresponse getjsonencoder awaitopenslotforrequest newdesc labelpairsorter closeremotechannels createcall action gc forresource samplecontroller flunders to do delete processraftrequest doappendinner isyesd verifydisksattachedtonode escape printnewline zero messageof isselected newmainkubelet startgarbagecollection monitorresizeevents launch startinterceptingoutput newproxier listenandservereadonly recordtosink runonce sendupdates listenandserve runhttphandler gethttprespbody handle head listpodstatsandupdatecpunanocoreusage listpodstats getcontainerstodeleteinpod updatestatusinternal newversion fromgrpc roundtrip handlereadloopresponse isliteral cpumask newstreamwatcher startinterceptingoutput tailfile startrequest announcespecruncompletion printfailure waitforchanges processraftrequest newwatchergrpcstream connect markdiskasdeletable parseproposectxerr writedatafromhandler writeframefromhandler sortstrings failuredetails endblock printspecfailure ioctlsetint registerevent withdeadline cancel emitscalar callk handleinterrupt configuretransports seterror addevent setbuffer constructblock listenforsignal volumesinuse volumesinusefunc unmarshal newparser unmarshalpointer warn compact isfeatureinrange errors insert remove do add pathstoremove add wait update reset flush transform send openexistingornew rotate header append copy new parse sync encode call new decode v set update getstartdiallocked normalizedname verify fillclusterformembers listengrpc msync newproxyserver createcall dialstreamlocal monitorsize launch createandinitkubelet specsuitewillbegin specsuitedidend hive newconn startkubelet newaggregator runonce newestablishingcontroller defaultcontrollerratelimiter fetchcontainerstats statistics hcsclosecomputesystem run startgarbagecollection sendupdates startinterceptingoutput createhealthcheck newtimeoutlistener listenandservereadonly monitorresizeevents startframewrite closenotify attachvolume portforward verifysignature eventf newsnapshotreadercloser runcmdcontext generateregisterpluginfunc createvolume processpodonnode runonce launch sendsnapshot newproxier handleconnection serve dispatch handlepodcleanups verify listenandserve removeandreadd syncservices lockedupdateloadbalancerhosts generateevent newmainkubelet mergeservice createdialer teardown unlock ref initservice handlecall decodearguments listenandserve serveudp newreflectcache store asfloat newunstructuredobjecttyper newsessionwithoptions newsession resolvenow init addextensiontype iterate update createmetadatas processsignalresult cpumask make truncatingformat callttbfunc calltrfunc maprange closeremotechannels newwithdelegate new format gracefulstop roundtripopt shouldretryrequest watcherrorstream mountserviceaccounttoken createvolume getfullqualifiedpluginnameforvolume setnodevolume mustnewmember listengrpc newloaderatconfirmeddir hasprefix create createuser inbounds run runonce listenandservereadonly handleconnection cleanuporphanedpodcgroups generateregisterpluginfunc listenandserve initforos addserviceonportinternal newmainkubelet startgarbagecollection monitorresizeevents newclientconnectionwithpings createvolume addserviceportportal sendupdates dispatch eventf processpodonnode verifysignature inbounds registerplugin startdbsizemonitorperendpoint handlepodcleanups newproxier recordtosink waitforattacheddevice addwork startcompactoronce mergeservice updatepod launch startinterceptingoutput dispatch verify removeandreadd syncservices chan attachvolume newsnapshotreadercloser execute receive rotate mill after unlock equalmessage equalunknown header fromcontext finish printf output serviceticket setthumbprint labelpairsorter doappendinner newstore newunsafeset migrateifneeded operationcompletehook setnodevolume start walkpaths reset protoreflect resetmessage createuser roundtrip startcompactoronce roundtripopt updatepod startdbsizemonitorperendpoint updateetcddbsize checkforupdates getifnewerthan hcsclosecomputesystem hcsclosecomputesystem lock ctx waitdeletes newerror validate newcrdfinalizer addeventhandler newdiscoveryrestmapper addspecific close send runraw runwithfriendlyerror valueencoder typeencoder addconnifneeded execincontainer inspectexec kill init waitforsingleobject receive parsenetlinkmessage newunstructuredobjecttyper waitcancelsubstreams getmetricsfromnode newsession definetests setupat migrateifneeded newfilepipeline newdecoratedwatcher newconn createmergedsnapshotmessage portforward updatetransport addpod buildauthn receive createcall evaluate specsuitedidend starttestserverordie sendmergedsnap startkubelet copyalllogs watchpods handleresponse startgatheringdata dgemm mountserviceaccounttoken passesverifyfilter copystdout rediscovernode copystderr watcherrorstream sink handleresizes createandinitkubelet freepages newstreamwatcher build scheduleframewrite dialwithconn newproxier newaggregator initializecsinode waitforavailableendpoint newclientconn newproxyserver listenforsignal applyentries copystdin inworker monitorsize launch dispatch runonce specsuitewillbegin dialstreamlocal createpd recvloop podsrunning verify applysnapshot sendloop eval getrequestedcontainers handlepodcleanups gc newlistener newtimeoutlistener syncpod containerstartspec log sweetenfields fetch getresponserange yaml emitter set unicode trimrightfunc get string getlatestpage errorf warn sub newparser yaml parser set input string recvkeepalive sendkeepaliveloop newproxyserver newproxier newfileset exec optionmanager mill startframewrite handlepodreconcile deletecontainersinpod markdiskasdeletable upload openfile getresponserange checktxnauth canset evaluate dispatch roundtrip roundtripopt newconn dialstreamlocal unregistercallback hcsunregistercomputesystemcallback kill errorf destroy yaml emitter delete serve shutdownin listenandserve listenandservekubeletserver appengineurlfetchhook finddefaultcredentials finddefaultcredentials newsnapshotreadercloser closenotify startframewrite portforward sendsnapshot detach dovolumeset addlivezchecks processpodonnode handleconnection syncservices handlepodcleanups mergeservice removeandreadd attachvolume generateregisterpluginfunc init compareandswappointer waitforavailableendpoint receive handleinterrupt updatesecuritygroup createnodesecuritygroup poperror withlabelvalues hcsunregistercomputesystemcallback hcsunregistercomputesystemcallback satisfiedexpectations poller deliversignal deliver removecontainer poperror cadvisorinfotoacceleratorstats getrequestedcontainers opencontainer fetchcontainerstats getcontainerstats requestlogger handleauth wordmask effectivelyequals fileline dialwithconn maybelastindex roundtripopt dispatch runstm newclientconnectionwithpings roundtrip drainedat removeimage sortbylength servergroupsandresources fetchgroupversionresources initforos initservice changeevents do errallfailed run newwatchergrpcstream bulkverifyvolumes initforos clientconfig newipvsrequest writelines starttestserver gather stringaffinecomparable dovolumesetgetresponse new getfiletype getfiletype addcertrotation pathprefix registerplugin applyserviceconfigandbalancer newccbalancerwrapper output objecttype sendconsumecustommetric syncpodfn wrapup listenandservepodresources now timefromunixnano applyentries listenandservereadonly listenandservekubeletreadonlyserver transformstartcontainererror objecttype dialstreamlocal newconn createpd createpd createpd createdisk selectloadbalancer getvmsetnames getprefix gather evaluate starttestserverordie dgemm markdiskasdeletable mountserviceaccounttoken processpodonnode applysnapshot handleresponse watcherrorstream createcall configuretransports verify newproxyserver newclientconn newlistener recvloop newconn receive addpeertoprober serveconn createandinitkubelet listenforsignal watchpods scheduleframewrite dispatch forresource addmember initializecsinode setupat run newaggregator startpeer createdialer createpd specsuitedidend mergeservice newcacherfromconfig resolvenow handlepodupdates createhealthcheck startkubelet removeandreadd podsrunning specsuitewillbegin rediscovernode runonce generateregisterpluginfunc newwithdelegate handleconnection monitorsize launch handleresizes getmetricsfromnode callk newproxier newdecoratedwatcher newclientconnectionwithpings portforward copystderr attachvolume generateevent deletecollection addpod sink newsession buildauthn sendloop build handlepodreconcile applyentries run newfilepipeline freepages copystdin copystdout passesverifyfilter applyserviceconfigandbalancer updatetransport recordtosink waitcancelsubstreams dialstreamlocal dispatch newdelayingqueuewithcustomclock handlepodcleanups createmergedsnapshotmessage migrateifneeded syncservices newstreamwatcher startgatheringdata waitforavailableendpoint copyalllogs definetests lockedupdateloadbalancerhosts newwatchergrpcstream inworker listenerroutine sendmergedsnap updateclientconnstate applyserviceconfigandbalancer switchbalancer newccbalancerwrapper checktxnreqspermission connect propsslice fatalf get log printf fataldepth logf header cancel fail exit newtimer now getintelrdtroot dispatch newclientconnectionwithpings updatetransport addcertrotation close eof newsession compileupdate parseserviceconfig newwatchergrpcstream podsrunning sendupdates sethttpheader string addmember httpmembers addmemberbyurl do debugresponse newcallset newfileset waitdeletes waitdelete getunknown createcall freepages inworker handlecall latestcontainerstats userdefinedmetrictype startcompactoronce startcompactor verifier newverifier marshal destroy newlimiter newdelayingqueuewithcustomclock tostringslice callk clockwait packkeywire keysfromremote node emitscalar format grpcstatus verifysignature verify sendmsg outpayload getfrompath nshandle warningln connectionerrorf print linesof strings string runhandler from nodesof from infof string new flushall add version lstat newunstructuredobjecttyper byprefixpriority newintset max sendsnap versionrange removechild configuretransports keysfromremote hcsunregistercomputesystemcallback stopdbsizemonitor typeencoder enable newnode splitprefix finish seterror parseserviceconfig sendconsumememrequests sendconsumecustommetric closenotify handlestreams sendconsumecpurequests newservertransport setup ioctlsetint run storescratch reset updateeffectivewindowsize deliver volumesinusefunc gethttprespbody run processdata optionmanager newdelayingqueue example dispatch newclientconnectionwithpings addlistener newdelayingqueue string grow chan walkoperation walkresponse header formatheader newctxclient withcancel runcmdcontext newcollector mapeventstocustomevents getifnewerthan injecthandlers pushfrontnamed setstate discovernode sortablememberslicebypeerurls handleauth get do retrieveunexportedfield unsafeaddr fdatasync msync newdelayingqueuewithcustomclock newdelayingqueue specsuitewillbegin escape setnodevolume verifydisksattachedtonode printspecfailure printfailure keytofilename sum unknown isunknown equalunknown attachvolume handleconnection newdelayingqueuewithcustomclock newwatchergrpcstream mergeservice removeandreadd configuretransports handlepodcleanups generateregisterpluginfunc recordtosink syncservices run processpodonnode configuretransports addconnifneeded createmergedsnapshotmessage newmessage public recordtosink scheduleframewrite iscontrol handlesetupsummary failuredetails announcesetupfailure verify dedup retrieve profile newwithtlsconfig proxyurl mustwrapvaluereflect runstm runonce newmux starttestserver sendconsumecpurequests handlestreams newsnapshotreadercloser newclientconnectionwithpings listenerroutine roundtripopt newwatchergrpcstream monitorresizeevents sendsnapshot newproxier chan closenotify handlepodcleanups startframewrite startnode discovernode newdelayingqueuewithcustomclock listenandservereadonly eventf addpeertoprober enable newmainkubelet sendupdates sendconsumememrequests processpodonnode verifysignature removeandreadd newclientconn deletecollection startgarbagecollection startpeer newsession sendconsumecustommetric dgemmparallel createpd updatetransport createvolume handleconnection startcompactoronce portforward updatepod startdbsizemonitorperendpoint restartnode newlessor mergeservice recordtosink run apply syncservices launch startinterceptingoutput dispatch keepaliveloop clienthandshake verify getcollector dispatch listenandserve attachvolume generateregisterpluginfunc handlesignal roundtrip createvolume apply panicf newreader newbpsloop terminatepod updatestatusinternal healthcheckfunc newclienttransport getloadbalancer getfloatingipbyportid add add withcancel cancel logname sendconsumecpurequests sendconsumememrequests sendconsumecustommetric closenotify handlestreams runstm newclusterauthenticationtrustcontroller keyfn clientconfig connect store retrieveone bytes panicmessage make dispatch newclientconnectionwithpings consumes refreshmuxlocked newmux newcond newmessage mapeventstocustomevents setup kill convertprocesserror openidkeysetjson publicjwksfromkeys dialstreamlocal createcall samplecontroller newdecoder convertslicenumbers withttl mkstm pushfrontnamed loadsources newfile action method stopnotify get reset trimrightfunc handleinterrupt configuretransports callk containerexecinspect newconn sendsnapshot sendsnap wait init waitforsingleobject parse reducefunc build enablecsm injecthandlers nodesof node grow decoderuneinstring withretries fetchgroupversionresources getworkerunsafe getwrappedworkerfunc fromgrpc withstack runwithfriendlyerror run rawsyscallnoerror remotecredprovider tailfilesync waitforchanges initialize rebuild installapigroups newaddress updateresolverstate profile difference newversionedset withlazyprefix freepages markdiskasdeletable connectionerrorf walkpathitem walkparameters unsafeaddr configuretransports upload zero canset settypeforkey newconditioncontroller addeventhandler stop new kill error removeendpoints do done close kill stop kill signal removecontainer wait string unlock newtimer lock finish after remove destroy v iterate iterateusing runstm read sum rescale rediscovernode roundtrip roundtripopt quo newconn dialstreamlocal freepages newwatchergrpcstream max max loadpackage newchecker remove wordmask buildauthn addeventhandlerwithresyncperiod addlistener newconnection sendpings getrequest newwatchergrpcstream connect keepalive deadlineloop watchpods marshal reflectvalue newclientconnectionwithpings newconnection bulkverifyvolumes setnodevolume syncpod syncpod size getlengthsid refreshmuxlocked byprefixpriority error errors singleretrieve retrieve dispatch removeinstance addinstances newwatchablestore syncvictimsloop readuvarint copystdin sum rawvector secondssincesshkeysync addlivezchecks portforward portforwardfunc wrapup checkforupdates new newloaderatconfirmeddir terminate monitor serveconn rediscovernode discovernode flunders readall readall constructblock numlinebreaks ctx delete waitdeletes updatestatusinternal updatelasttransitiontime put reset reset remotecredprovider executecredentialprocess createdialer createhealthcheck passesverifyfilter verify sync observe fn write send printwithfileline format protoreflect importlayer layerpath restartnode evictoldest newwatchergrpcstream deletecollection newclienttransport reducefunc offset id timefromunixnano isyesd newclientconnectionwithpings dispatch newcollector mapeventstocustomevents createnodesecuritygroup toruleprotocol commit unsafecommit evict evictoldest destroy getpath timeoutio configuretransports enable roundtripopt startrequest resolvenow newclientconnectionwithpings roundtrip applyserviceconfigandbalancer gather newwatchergrpcstream dispatch getdelegate newdiscoveryrestmapper handlepodupdates dispatchwork delims roundtrip roundtripopt decoderuneinstring errorf getunknown seterror gettrimmed timeoutio quotechar dispatch keysfromremote send sendmsg inf closeremotechannels revoke values pingtimeoutseconds configuretransports dispatchwork terminatepod copystdout unmarshalpointer initializecsinode getdetail cancelrequest newclientconn newmux setupat totalsecretbytes marshaldoc nodev newleasefromleaseclient withcancel getwithcontext singleretrieve specdidcomplete failuredetails listenandservekubeletreadonlyserver listenandserve create logname keepaliveloop replytoclient parse newkey withlabelvalues getmetricwithlabelvalues log write build cadvisorinfotonetworkstats latestcontainerstats warningln exit print mapindex equalmessage appenduint dispatch newclientconnectionwithpings flushall sync leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 1 |
752,141 | 26,274,848,453 | IssuesEvent | 2023-01-06 20:48:07 | AngelUMICH/Long-Time-Dynamics | https://api.github.com/repos/AngelUMICH/Long-Time-Dynamics | closed | energy limit algorithm | New Feature priority | ## Details
Energy limit / threshold can be set automatically by determining the energy that will invoke a jump in KMC time that is equal to or less than the dpa rate or the rate at which the defects are introduced into the system. This energy limit can be calculated using rate theory. An equation should be included into the code in the INIT.f file.
## Definition of Done
- [x] Determine the equation for the energy limit.
- [x] Post the equation as a comment below
- [x] Energy limit is calculated at the start of each simulation | 1.0 | energy limit algorithm - ## Details
Energy limit / threshold can be set automatically by determining the energy that will invoke a jump in KMC time that is equal to or less than the dpa rate or the rate at which the defects are introduced into the system. This energy limit can be calculated using rate theory. An equation should be included into the code in the INIT.f file.
## Definition of Done
- [x] Determine the equation for the energy limit.
- [x] Post the equation as a comment below
- [x] Energy limit is calculated at the start of each simulation | non_test | energy limit algorithm details energy limit threshold can be set automatically by determining the energy that will invoke a jump in kmc time that is equal to or less than the dpa rate or the rate at which the defects are introduced into the system this energy limit can be calculated using rate theory an equation should be included into the code in the init f file definition of done determine the equation for the energy limit post the equation as a comment below energy limit is calculated at the start of each simulation | 0 |
31,784 | 7,451,966,085 | IssuesEvent | 2018-03-29 06:22:52 | w4824112w/ywgk | https://api.github.com/repos/w4824112w/ywgk | opened | 业务新增的接口,有待优化,入参可用对象接收 | code review | 以下这个接口中,新增的业务,提交的参数比较多的情况下,可用对象来接收,且不需要再进行单个单个的去set值了。
/**
* 新增家属注册信息
*
* @param request
* @param response
* @param name
* @param phone
* @param gender
* @param uuid
* @param jailId
* @param prisonerNumber
* @param relationship
* @return
*/
@RequestMapping(value = "/add", method = RequestMethod.POST)
public JSONObject add(HttpServletRequest request,
HttpServletResponse response, String familyId, String name,
String phone, String gender, String uuid, Integer jailId,
String prisonerNumber, String relationship, String prisonerId) {
Registrations registrations = new Registrations();
registrations.setFamilyId(Integer.parseInt(familyId));
registrations.setName(name);
registrations.setPhone(phone);
registrations.setGender(gender);
registrations.setUuid(uuid);
registrations.setJailId(jailId);
registrations.setPrisonerNumber(prisonerNumber);
registrations.setRelationship(relationship);
registrations.setPrisonerId(Integer.parseInt(prisonerId));
registrations.setCreatedAt(new Date());
registrations.setUpdatedAt(new Date());
registrations = registrationsService.save(registrations);
JSONObject result = new JSONObject();
result.put("registrations", registrations);
log.info("新增家属注册信息接口成功");
return ErrorEnums.getResult(ErrorEnums.SUCCESS, "新增家属注册信息", result);
} | 1.0 | 业务新增的接口,有待优化,入参可用对象接收 - 以下这个接口中,新增的业务,提交的参数比较多的情况下,可用对象来接收,且不需要再进行单个单个的去set值了。
/**
* 新增家属注册信息
*
* @param request
* @param response
* @param name
* @param phone
* @param gender
* @param uuid
* @param jailId
* @param prisonerNumber
* @param relationship
* @return
*/
@RequestMapping(value = "/add", method = RequestMethod.POST)
public JSONObject add(HttpServletRequest request,
HttpServletResponse response, String familyId, String name,
String phone, String gender, String uuid, Integer jailId,
String prisonerNumber, String relationship, String prisonerId) {
Registrations registrations = new Registrations();
registrations.setFamilyId(Integer.parseInt(familyId));
registrations.setName(name);
registrations.setPhone(phone);
registrations.setGender(gender);
registrations.setUuid(uuid);
registrations.setJailId(jailId);
registrations.setPrisonerNumber(prisonerNumber);
registrations.setRelationship(relationship);
registrations.setPrisonerId(Integer.parseInt(prisonerId));
registrations.setCreatedAt(new Date());
registrations.setUpdatedAt(new Date());
registrations = registrationsService.save(registrations);
JSONObject result = new JSONObject();
result.put("registrations", registrations);
log.info("新增家属注册信息接口成功");
return ErrorEnums.getResult(ErrorEnums.SUCCESS, "新增家属注册信息", result);
} | non_test | 业务新增的接口,有待优化,入参可用对象接收 以下这个接口中,新增的业务,提交的参数比较多的情况下,可用对象来接收,且不需要再进行单个单个的去set值了。 新增家属注册信息 param request param response param name param phone param gender param uuid param jailid param prisonernumber param relationship return requestmapping value add method requestmethod post public jsonobject add httpservletrequest request httpservletresponse response string familyid string name string phone string gender string uuid integer jailid string prisonernumber string relationship string prisonerid registrations registrations new registrations registrations setfamilyid integer parseint familyid registrations setname name registrations setphone phone registrations setgender gender registrations setuuid uuid registrations setjailid jailid registrations setprisonernumber prisonernumber registrations setrelationship relationship registrations setprisonerid integer parseint prisonerid registrations setcreatedat new date registrations setupdatedat new date registrations registrationsservice save registrations jsonobject result new jsonobject result put registrations registrations log info 新增家属注册信息接口成功 return errorenums getresult errorenums success 新增家属注册信息 result | 0 |
288,329 | 24,898,760,009 | IssuesEvent | 2022-10-28 18:29:26 | hashicorp/terraform-provider-google | https://api.github.com/repos/hashicorp/terraform-provider-google | opened | Failing test(s): TestAccDNSManagedZone_dnsManagedZoneQuickstartExample | test failure | <!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. --->
<!-- i.e. "Consistently since X date" or "X% failure in MONTH" -->
Failure rate: 100% since 2022-09-24
<!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. -->
Impacted tests:
- TestAccDNSManagedZone_dnsManagedZoneQuickstartExample
<!-- Link to the nightly build(s), ideally with one impacted test opened -->
Nightly builds:
- https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloud_ProviderGoogleCloudGoogleProject/348525?buildTab=tests&expandedTest=-3604989738191621873
<!-- The error message that displays in the tests tab, for reference -->
Message:
```
Error: Error creating ManagedZone: googleapi: Error 400: Please verify ownership of the 'googlecloudexample.com.' domain (or a parent) at http://www.google.com/webmasters/verification/ and try again
```
Similar issue: https://github.com/hashicorp/terraform-provider-google/issues/12528
Previously this was failing with this error sporadically - possibly this was a new verification that rolled out? | 1.0 | Failing test(s): TestAccDNSManagedZone_dnsManagedZoneQuickstartExample - <!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. --->
<!-- i.e. "Consistently since X date" or "X% failure in MONTH" -->
Failure rate: 100% since 2022-09-24
<!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. -->
Impacted tests:
- TestAccDNSManagedZone_dnsManagedZoneQuickstartExample
<!-- Link to the nightly build(s), ideally with one impacted test opened -->
Nightly builds:
- https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloud_ProviderGoogleCloudGoogleProject/348525?buildTab=tests&expandedTest=-3604989738191621873
<!-- The error message that displays in the tests tab, for reference -->
Message:
```
Error: Error creating ManagedZone: googleapi: Error 400: Please verify ownership of the 'googlecloudexample.com.' domain (or a parent) at http://www.google.com/webmasters/verification/ and try again
```
Similar issue: https://github.com/hashicorp/terraform-provider-google/issues/12528
Previously this was failing with this error sporadically - possibly this was a new verification that rolled out? | test | failing test s testaccdnsmanagedzone dnsmanagedzonequickstartexample failure rate since impacted tests testaccdnsmanagedzone dnsmanagedzonequickstartexample nightly builds message error error creating managedzone googleapi error please verify ownership of the googlecloudexample com domain or a parent at and try again similar issue previously this was failing with this error sporadically possibly this was a new verification that rolled out | 1 |
42,392 | 10,965,863,165 | IssuesEvent | 2019-11-28 04:50:24 | joelday/papyrus-lang | https://api.github.com/repos/joelday/papyrus-lang | closed | Problem with pyro compiling individually listed scripts? | bug pyro/build | Filed as https://github.com/fireundubh/pyro/issues/3
**Describe the bug**
JonathanOstrusToday at 2:36 AM:
```
(VSCode papyrus-lang fork 1.3.3vsc.77)
[CMPL] Compiling "TestScript"...
[WARN] No source scripts were indexed. Possible reason: No source scripts were recently modified.
Traceback (most recent call last):
File "c:\Users\Dream Machine\.vscode\extensions\joelday.papyrus-lang-vscode-2.19.1\pyro\pyro_cli\__main__.py", line 145, in <module>
File "c:\Users\Dream Machine\.vscode\extensions\joelday.papyrus-lang-vscode-2.19.1\pyro\pyro_cli\__main__.py", line 81, in run
ValueError: not enough values to unpack (expected 2, got 0)
The terminal process terminated with exit code: 1
```
It actually compiled the file tho
this is the ppj
```
<?xml version='1.0'?><!-- Game is either sse, tesv, or fo4 -->
<PapyrusProject xmlns="PapyrusProject.xsd"
Flags="TESV_Papyrus_Flags.flg"
Game="sse"
Output="Scripts"
Optimize="false"
Release="false"
Final="false">
<Imports>
<Import>.\Source\Scripts</Import>
<Import>I:\Games\Steam Library\steamapps\common\Skyrim Special Edition\Data\Source\Scripts</Import>
</Imports>
<Scripts>
<Script>TestScript</Script>
</Scripts>
</PapyrusProject>
```
**To Reproduce**
Steps to reproduce the behavior:
See previous.
| 1.0 | Problem with pyro compiling individually listed scripts? - Filed as https://github.com/fireundubh/pyro/issues/3
**Describe the bug**
JonathanOstrusToday at 2:36 AM:
```
(VSCode papyrus-lang fork 1.3.3vsc.77)
[CMPL] Compiling "TestScript"...
[WARN] No source scripts were indexed. Possible reason: No source scripts were recently modified.
Traceback (most recent call last):
File "c:\Users\Dream Machine\.vscode\extensions\joelday.papyrus-lang-vscode-2.19.1\pyro\pyro_cli\__main__.py", line 145, in <module>
File "c:\Users\Dream Machine\.vscode\extensions\joelday.papyrus-lang-vscode-2.19.1\pyro\pyro_cli\__main__.py", line 81, in run
ValueError: not enough values to unpack (expected 2, got 0)
The terminal process terminated with exit code: 1
```
It actually compiled the file tho
this is the ppj
```
<?xml version='1.0'?><!-- Game is either sse, tesv, or fo4 -->
<PapyrusProject xmlns="PapyrusProject.xsd"
Flags="TESV_Papyrus_Flags.flg"
Game="sse"
Output="Scripts"
Optimize="false"
Release="false"
Final="false">
<Imports>
<Import>.\Source\Scripts</Import>
<Import>I:\Games\Steam Library\steamapps\common\Skyrim Special Edition\Data\Source\Scripts</Import>
</Imports>
<Scripts>
<Script>TestScript</Script>
</Scripts>
</PapyrusProject>
```
**To Reproduce**
Steps to reproduce the behavior:
See previous.
| non_test | problem with pyro compiling individually listed scripts filed as describe the bug jonathanostrustoday at am vscode papyrus lang fork compiling testscript no source scripts were indexed possible reason no source scripts were recently modified traceback most recent call last file c users dream machine vscode extensions joelday papyrus lang vscode pyro pyro cli main py line in file c users dream machine vscode extensions joelday papyrus lang vscode pyro pyro cli main py line in run valueerror not enough values to unpack expected got the terminal process terminated with exit code it actually compiled the file tho this is the ppj papyrusproject xmlns papyrusproject xsd flags tesv papyrus flags flg game sse output scripts optimize false release false final false source scripts i games steam library steamapps common skyrim special edition data source scripts testscript to reproduce steps to reproduce the behavior see previous | 0 |
401,368 | 11,789,216,558 | IssuesEvent | 2020-03-17 16:47:10 | dgraph-io/dgraph | https://api.github.com/repos/dgraph-io/dgraph | closed | Binary Backups - AWS temporary credentials don't work | area/commercial/backup priority/P1 status/accepted | <!-- If you suspect this could be a bug, follow the template. -->
### What version of Dgraph are you using?
1.2.0
### Have you tried reproducing the issue with the latest release?
Only 1.2.0
### What is the hardware spec (RAM, OS)?
EC2 instance
### Steps to reproduce the issue (command/config used to run Dgraph).
Running dgraph with docker-compose
1) Do not set any AWS_* envvars inside the docker envs for the zero or alpha
2) perform a curl `curl -XPOST localhost:8080/admin/backup -d "destination=s3:///<bucketname>"` but also pass as parameters the AWS temporary credentials for The access_key, secret_key, and session_token.
### Expected behaviour and actual result.
The backup should be successfully completed, actual result the backup fails with permission failure.
Not clear to me is:
1) If you do not have static AWS credentials how does this work?
2) Do you have to have AWS* envvars set in order to override any of those values in first place?
3) Can you use temporary AWS credentials?
| 1.0 | Binary Backups - AWS temporary credentials don't work - <!-- If you suspect this could be a bug, follow the template. -->
### What version of Dgraph are you using?
1.2.0
### Have you tried reproducing the issue with the latest release?
Only 1.2.0
### What is the hardware spec (RAM, OS)?
EC2 instance
### Steps to reproduce the issue (command/config used to run Dgraph).
Running dgraph with docker-compose
1) Do not set any AWS_* envvars inside the docker envs for the zero or alpha
2) perform a curl `curl -XPOST localhost:8080/admin/backup -d "destination=s3:///<bucketname>"` but also pass as parameters the AWS temporary credentials for The access_key, secret_key, and session_token.
### Expected behaviour and actual result.
The backup should be successfully completed, actual result the backup fails with permission failure.
Not clear to me is:
1) If you do not have static AWS credentials how does this work?
2) Do you have to have AWS* envvars set in order to override any of those values in first place?
3) Can you use temporary AWS credentials?
| non_test | binary backups aws temporary credentials don t work what version of dgraph are you using have you tried reproducing the issue with the latest release only what is the hardware spec ram os instance steps to reproduce the issue command config used to run dgraph running dgraph with docker compose do not set any aws envvars inside the docker envs for the zero or alpha perform a curl curl xpost localhost admin backup d destination but also pass as parameters the aws temporary credentials for the access key secret key and session token expected behaviour and actual result the backup should be successfully completed actual result the backup fails with permission failure not clear to me is if you do not have static aws credentials how does this work do you have to have aws envvars set in order to override any of those values in first place can you use temporary aws credentials | 0 |
278,291 | 24,143,249,185 | IssuesEvent | 2022-09-21 16:24:36 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | opened | Manual test run on Linux for 1.43.x - Release #6 | tests OS/Linux QA/Yes release-notes/exclude OS/Desktop | ### Installer
- [ ] Check the installer is close to the size of the last release
- [ ] Check signature:
- [ ] If macOS, using x64 binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted`
- [ ] If macOS, using universal binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted`
- [ ] If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### About pages
- [ ] Verify that both `chrome://` and `about://` forward to `brave://` (run through several internal pages)
### Importing
- [ ] Verify that you can import `History`, `Favorites/Bookmarks` and `Passwords` from Google Chrome
- [ ] Verify that you can import `History`, `Favorites/Bookmarks`, `Passwords`, `Search Engines` and `Autofill` from Firefox
- [ ] Verify that you can import `Favorites/Bookmarks` from Microsoft Edge
- [ ] Verify that importing bookmarks using `Bookmark HTML File` retains the folder structure on a clean profile
### Context menus
- [ ] Verify you can block a page element using `Block element via selector` context-menu item
- [ ] Verify selecting `Manage custom filters` opens `brave://adblock` in a NTP
- [ ] Verify removing the rule from `brave://adblock` reflects the change on the website, after reload
### Extensions/Plugins
- [ ] Verify pdfium, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled (they don't show up in `brave://extensions`)
- [ ] Verify older version of an extension gets updated to new version via Google server
- [ ] Verify that `magnet` links and `.torrent` files correctly open WebTorrent and you're able to download the file(s)
- **Tip:** Free torrents available via https://webtorrent.io/free-torrents
### Chrome Web Store (CWS)
- [ ] Verify that installing https://chrome.google.com/webstore/detail/adblock-plus-free-ad-bloc/cfhdojbkjhnklbpkdaibdccddilifddb from CWS displays the `Brave has not reviewed the extension.` warning via the "Add Extension" modal
- [ ] Verify that installing https://chrome.google.com/webstore/detail/lastpass-free-password-ma/hdokiejnpimakedhajhdlcegeplioahd from CWS doesn't display the `Brave has not reviewed the extension.` warning via the "Add Extension" modal
### PDF
- [ ] Test that you can print a PDF
- [ ] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf
- [ ] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf
- [ ] Test that https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.105.6357&rep=rep1&type=pdf opens without issues
### Widevine
- [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine
- [ ] Verify `Widevine Notification` is shown when you visit HBO Max for the first time
- [ ] Test that you can stream on HBO Max on a fresh profile after installing Widevine
- [ ] If macOS, run the above Widevine tests for both `x64` and `universal` builds
### Geolocation
- [ ] Check that https://browserleaks.com/geo works and shows correct location
- [ ] Check that https://html5demos.com/geo/ works but doesn't require an accurate location
### Crash Reporting
- [ ] Check that loading `brave://crash` & `brave://gpucrash` causes the new tab to crash
- [ ] Check that `brave://crashes` lists the `Uploaded Crash Report ID` once the report has been submitted
- [ ] Verify the crash ID matches the report on Backtrace using `_rxid equal [ value ]`
### Bravery settings
- [ ] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [ ] Turning HTTPS Everywhere off and Shields off both disable the redirect to https://https-everywhere.badssl.com/
- [ ] Verify that toggling `Trackers & ads blocked` works as expected
- [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [ ] Visit https://www.blizzard.com and then turn on script blocking, page should not load.
- [ ] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
- [ ] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run)
- [ ] In `brave://settings/security`, choose a DNS provider from the providers listed under Use secure DNS, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown; only your chosen DoH provider should appear.
- [ ] Open a New Private Window with Tor, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown.
### TLS Pinning
- [ ] Visit https://ssl-pinning.someblog.org/ and verify a pinning error is displayed
- [ ] Visit https://pinning-test.badssl.com/ and verify a pinning error is **not** displayed
### Fingerprint Tests
- [ ] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address for each option under `Settings -> Privacy and Security -> WebRTC IP handling policy`
### Brave Ads
- [ ] Verify when you enable Rewards from panel or `brave://rewards`, Ads are enabled by default
- [ ] Verify Ads UI (panel, settings, etc) shows when in a region with Ads support
- [ ] Verify Ads UI (panel, settings, etc) does not show when in a region without Ads support. Verify the Ads panel does show the 'Sorry! Ads are not yet available in your region.' message.
- [ ] Verify when the system language is English, the Browser language is French, and you are in one of the supported regions, Ad notifications are still served to you.
- [ ] Verify you are served Ad notifications when Ads are enabled
- [ ] Verify ad earnings are reflected in the rewards widget on the NTP.
- [ ] Verify when Ads are toggled off, there are no Ad messages in the logs
- [ ] Verify when Rewards are toggled off (but Ads were not explicitly toggled off), there are no Ads logs recorded
- [ ] Verify view/click/dismiss/landed ad notifications show in `confirmations.json`
- [ ] Verify pages you browse to are being classified in the logs
- [ ] Verify tokens are redeemed by viewing the logs (you can use `--rewards=debug=true` to shorten redemption time)
- [ ] Verify Ad is not shown if a tab is playing media and is only shown after it stops playing
### Rewards
- [ ] Verify that none of the reward endpoints are being contacted when a user visits a media publisher (`youtube.com`, `reddit.com`, `twitter.com`, `github.com`) and hasn't interacted with rewards
- [ ] Verify that `rewards.brave.com`, `pcdn.brave.com`, `grant.rewards.brave.com` or `api.rewards.brave.com` are not being contacted
- [ ] Verify you are able to create a new Rewards profile.
- [ ] Verify you are able to restore an old Rewards profile.
- [ ] Verify Rewards balance shows correct BAT and USD value.
- [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in panel transactions list.
- [ ] Verify AC monthly budget shows correct BAT and USD value.
- [ ] Verify you are able to exclude a publisher from the auto-contribute table and popup list of sites.
- [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel.
- [ ] Verify you are able to perform an auto contribution.
- [ ] Verify auto contribution is reflected in the rewards widget on the new-tab page (NTP).
- [ ] Verify monthly statement shows expected data.
- [ ] Verify when you click on the BR panel while on a site, the panel displays site-specific information (site favicon, domain, attention %).
- [ ] Verify BR panel shows message about an unverified publisher.
- [ ] Verify one time and monthly tip banners show a message about unverified publisher.
- [ ] Verify one time tip and monthly tip banners show a verified checkmark for a verified creator.
- [ ] Verify when you click on `Send a tip`, the custom tip banner displays if set up.
- [ ] Verify custom tip banner is also displayed for monthly contribution.
- [ ] Verify you are able to make one-time tip and they display in Tips panel.
- [ ] Verify tip is reflected in the rewards widget on the NTP.
- [ ] Verify when you tip an unverified publisher, the one time tip is recorded in the Pending Contributions list.
- [ ] Verify you are able to make recurring tip and they display in Monthly Contributions panel.
- [ ] Verify you are able to adjust your recurring tip amount from the BR panel.
- [ ] Verify recurring tip is reflected in the rewards widget on the NTP.
- [ ] Verify you can tip a verified website.
- [ ] Verify the website displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified YouTube creator.
- [ ] Verify the YouTube creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Vimeo creator.
- [ ] Verify the Vimeo creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Twitch creator.
- [ ] Verify the Twitch creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Twitter user from the panel.
- [ ] Verify you can tip a verified Twitter user via inline tip button.
- [ ] Verify the in-line tip button is spaced properly.
- [ ] Verify you can tip a verified GitHub user from the panel.
- [ ] Verify you can tip a verified GitHub user via inline tip button.
- [ ] Verify the GitHub creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Reddit user from the panel.
- [ ] Verify you can tip a verified Reddit user via inline tip button.
- [ ] Verify if you disable auto-contribute you are still able to tip creators.
- [ ] Verify if auto-contribute is disabled AC does not occur.
- [ ] Verify if Rewards is disabled AC does not occur.
- [ ] Verify that disabling Rewards and enabling it again does not lose state.
- [ ] Adjust min visit/time in settings. Visit some sites to verify they are added to the table after the specified settings.
- [ ] Uphold cases
- [ ] Verify you are able to connect a KYC'd Uphold account to Rewards.
- [ ] Verify balance in Brave updates when BAT is added to the Brave Browser card.
- [ ] Verify if you only have user-controlled BAT (BAT in Uphold only), you can only tip KYC'd creators, any tips to non-KYC'd creators go to the Pending Contributions list.
- [ ] Verify connected (verified but not KYC'd) publishers display messaging on panel and tip banner.
- [ ] Verify you are able to perform an auto contribute using Uphold BAT.
### Social-media blocking settings
- [ ] Verify individual `Social media blocking` buttons works as intended when enabled/disabled by visiting https://fmarier.github.io/brave-testing/social-widgets.html
- [ ] visit `brave://settings/privacy` -> `Site and Shields Settings` -> `Cookies and site data` and ensure that
- [ ] both `https://[*.]firebaseapp.com` & `https://accounts.google.com` are added into `Sites that can always use cookies` when `Allow Google login buttons on third party sites` is enabled
- [ ] both `https://[*.]firebaseapp.com` & `https://accounts.google.com` are removed from `Sites that can always use cookies` when `Allow Google login buttons on third party sites` is disabled
- [ ] ensure that you can log in into https://www.expensify.com while `Allow Google login buttons on third party sites` is enabled
- [ ] ensure that once `Allow Google login buttons on third party sites` has been disabled, you can't log in into https://www.expensify.com
### Sync
- [ ] Verify you are able to create a sync chain and add a mobile/computer to the chain
- [ ] Verify you are able to join an existing sync chain using code words
- [ ] Verify the device name is shown properly when sync chain is created
- [ ] Verify you are able to add a new mobile device to the chain via QR code/code words
- [ ] Verify newly created bookmarks get sync'd to all devices on the sync chain
- [ ] Verify existing bookmarks on current profile gets sync'd to all devices on the sync chain
- [ ] Verify folder structure is retained after sync completes
- [ ] Verify bookmarks don't duplicate when sync'd from other devices
- [ ] Verify removing bookmark from device gets sync'd to all devices on the sync chain
- [ ] Verify adding/removing a bookmark in offline mode gets sync'd to all devices on the sync chain when device comes online
- [ ] With only two devices in chain, verify removing the other device resets the sync on b-c as well
### Tor Tabs
- [ ] Visit https://check.torproject.org in a Tor window, ensure it shows a success message for using a Tor exit node
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Click `New Tor connection for this site` in app menu, ensure the exit node IP address changes after page is reloaded
- [ ] Visit https://protonmailrmez3lotccipshtkleegetolb73fuirgj7r4o4vfu7ozyd.onion, https://brave4u7jddbv7cyviptqjc7jusxh72uik7zt6adtckl5f4nwy2v72qd.onion/, and https://search.brave4u7jddbv7cyviptqjc7jusxh72uik7zt6adtckl5f4nwy2v72qd.onion/ in a Tor window and ensure all pages resolve
- [ ] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown
- [ ] Verify Torrent viewer doesn't load in a Tor window
- [ ] Ensure you are able to download a file in a Tor window. Verify all Download/Cancel, Download/Retry and Download works in Tor window
### Cookie and Cache
- [ ] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the evercookie site does not remember the old evercookie value
### Chromium/Brave GPU
- [ ] Verify that `brave://gpu` (Brave) matches `chrome://gpu` (Chrome) when using the same Chromium version
### Startup & Components
- [ ] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler`, `Wireshark` or `LittleSnitch` (or a similar application)
- [ ] Verify that opening a NTP doesn't trigger any outbound connections related to widgets without user interaction
- [ ] Remove the following component folders and ensure that they're being re-downloaded after restarting the browser:
- [ ] `afalakplffnnnlkncjhbmahjfjhmlkal`: `AutoplayWhitelist.dat`, `ExtensionWhitelist.dat`, `ReferrerWhitelist.json` and `Greaselion.json`
- [ ] `CertificateRevocation`
- [ ] `cffkpbalmllkdoenhmdmpbkajipdjfam`: `rs-ABPFilterParserData.dat` & `regional_catalog.json` (AdBlock)
- [ ] `gccbbckogglekeggclmmekihdgdpdgoe`: (Sponsored New Tab Images)
- [ ] `jicbkmdloagakknpihibphagfckhjdih`: `speedreader-updater.dat`
- [ ] `oofiananboodjbbmdelgdommihjbkfag`: HTTPSE
- [ ] `Safe Browsing`
- [ ] Restart the browser, load `brave://components`, wait for 8 mins and verify that no component shows any errors
**Note:** Always double check `brave://components` to make sure there's no errors/missing version numbers
### Session storage
- [ ] Temporarily move away your browser profile and test that a new profile is created on browser relaunch
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [ ] Test that both windows and tabs are being restored, including the current active tab
- [ ] Ensure that tabs are being lazy loaded when a previous session is being restored
### Upgrade
- [ ] Make sure that data from the last version appears in the new version OK
- [ ] Ensure that `brave://version` lists the expected Brave & Chromium versions
- [ ] With data from the last version, verify that:
- [ ] Bookmarks on the bookmark toolbar and bookmark folders can be opened
- [ ] Cookies are preserved
- [ ] Installed extensions are retained and work correctly
- [ ] Opened tabs can be reloaded
- [ ] Stored passwords are preserved
- [ ] Sync chain created in previous version is retained
- [ ] Social media-blocking buttons changes are retained
- [ ] Rewards
- [ ] BAT balance is retained
- [ ] Auto-contribute list is retained
- [ ] Both Tips and Monthly Contributions are retained
- [ ] Panel transactions list is retained
- [ ] Changes to rewards settings are retained
- [ ] Ensure that Auto Contribute is not being enabled when upgrading to a new version if AC was disabled
- [ ] Ads
- [ ] Both `Estimated pending rewards` & `Ad notifications received this month` are retained
- [ ] Changes to ads settings are retained
- [ ] Ensure that ads are not being enabled when upgrading to a new version if they were disabled
- [ ] Ensure that ads are not disabled when upgrading to a new version if they were enabled
## Hardware Wallet Test (To be checked on each major CR bump)
- [ ] Verify you can link hardware wallet using Trezor and unlock the wallet
- [ ] Verify you can link hardware wallet using Ledger Nano and unlock the wallet
- [ ] Verify you can perform a transaction with hardware wallet using Ledger Nano or Trezor
| 1.0 | Manual test run on Linux for 1.43.x - Release #6 - ### Installer
- [ ] Check the installer is close to the size of the last release
- [ ] Check signature:
- [ ] If macOS, using x64 binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted`
- [ ] If macOS, using universal binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted`
- [ ] If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### About pages
- [ ] Verify that both `chrome://` and `about://` forward to `brave://` (run through several internal pages)
### Importing
- [ ] Verify that you can import `History`, `Favorites/Bookmarks` and `Passwords` from Google Chrome
- [ ] Verify that you can import `History`, `Favorites/Bookmarks`, `Passwords`, `Search Engines` and `Autofill` from Firefox
- [ ] Verify that you can import `Favorites/Bookmarks` from Microsoft Edge
- [ ] Verify that importing bookmarks using `Bookmark HTML File` retains the folder structure on a clean profile
### Context menus
- [ ] Verify you can block a page element using `Block element via selector` context-menu item
- [ ] Verify selecting `Manage custom filters` opens `brave://adblock` in a NTP
- [ ] Verify removing the rule from `brave://adblock` reflects the change on the website, after reload
### Extensions/Plugins
- [ ] Verify pdfium, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled (they don't show up in `brave://extensions`)
- [ ] Verify older version of an extension gets updated to new version via Google server
- [ ] Verify that `magnet` links and `.torrent` files correctly open WebTorrent and you're able to download the file(s)
- **Tip:** Free torrents available via https://webtorrent.io/free-torrents
### Chrome Web Store (CWS)
- [ ] Verify that installing https://chrome.google.com/webstore/detail/adblock-plus-free-ad-bloc/cfhdojbkjhnklbpkdaibdccddilifddb from CWS displays the `Brave has not reviewed the extension.` warning via the "Add Extension" modal
- [ ] Verify that installing https://chrome.google.com/webstore/detail/lastpass-free-password-ma/hdokiejnpimakedhajhdlcegeplioahd from CWS doesn't display the `Brave has not reviewed the extension.` warning via the "Add Extension" modal
### PDF
- [ ] Test that you can print a PDF
- [ ] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf
- [ ] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf
- [ ] Test that https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.105.6357&rep=rep1&type=pdf opens without issues
### Widevine
- [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine
- [ ] Verify `Widevine Notification` is shown when you visit HBO Max for the first time
- [ ] Test that you can stream on HBO Max on a fresh profile after installing Widevine
- [ ] If macOS, run the above Widevine tests for both `x64` and `universal` builds
### Geolocation
- [ ] Check that https://browserleaks.com/geo works and shows correct location
- [ ] Check that https://html5demos.com/geo/ works but doesn't require an accurate location
### Crash Reporting
- [ ] Check that loading `brave://crash` & `brave://gpucrash` causes the new tab to crash
- [ ] Check that `brave://crashes` lists the `Uploaded Crash Report ID` once the report has been submitted
- [ ] Verify the crash ID matches the report on Backtrace using `_rxid equal [ value ]`
### Bravery settings
- [ ] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [ ] Turning HTTPS Everywhere off and Shields off both disable the redirect to https://https-everywhere.badssl.com/
- [ ] Verify that toggling `Trackers & ads blocked` works as expected
- [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [ ] Visit https://www.blizzard.com and then turn on script blocking, page should not load.
- [ ] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
- [ ] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run)
- [ ] In `brave://settings/security`, choose a DNS provider from the providers listed under Use secure DNS, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown; only your chosen DoH provider should appear.
- [ ] Open a New Private Window with Tor, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown.
### TLS Pinning
- [ ] Visit https://ssl-pinning.someblog.org/ and verify a pinning error is displayed
- [ ] Visit https://pinning-test.badssl.com/ and verify a pinning error is **not** displayed
### Fingerprint Tests
- [ ] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address for each option under `Settings -> Privacy and Security -> WebRTC IP handling policy`
### Brave Ads
- [ ] Verify when you enable Rewards from panel or `brave://rewards`, Ads are enabled by default
- [ ] Verify Ads UI (panel, settings, etc) shows when in a region with Ads support
- [ ] Verify Ads UI (panel, settings, etc) does not show when in a region without Ads support. Verify the Ads panel does show the 'Sorry! Ads are not yet available in your region.' message.
- [ ] Verify when the system language is English, the Browser language is French, and you are in one of the supported regions, Ad notifications are still served to you.
- [ ] Verify you are served Ad notifications when Ads are enabled
- [ ] Verify ad earnings are reflected in the rewards widget on the NTP.
- [ ] Verify when Ads are toggled off, there are no Ad messages in the logs
- [ ] Verify when Rewards are toggled off (but Ads were not explicitly toggled off), there are no Ads logs recorded
- [ ] Verify view/click/dismiss/landed ad notifications show in `confirmations.json`
- [ ] Verify pages you browse to are being classified in the logs
- [ ] Verify tokens are redeemed by viewing the logs (you can use `--rewards=debug=true` to shorten redemption time)
- [ ] Verify Ad is not shown if a tab is playing media and is only shown after it stops playing
### Rewards
- [ ] Verify that none of the reward endpoints are being contacted when a user visits a media publisher (`youtube.com`, `reddit.com`, `twitter.com`, `github.com`) and hasn't interacted with rewards
- [ ] Verify that `rewards.brave.com`, `pcdn.brave.com`, `grant.rewards.brave.com` or `api.rewards.brave.com` are not being contacted
- [ ] Verify you are able to create a new Rewards profile.
- [ ] Verify you are able to restore an old Rewards profile.
- [ ] Verify Rewards balance shows correct BAT and USD value.
- [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in panel transactions list.
- [ ] Verify AC monthly budget shows correct BAT and USD value.
- [ ] Verify you are able to exclude a publisher from the auto-contribute table and popup list of sites.
- [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel.
- [ ] Verify you are able to perform an auto contribution.
- [ ] Verify auto contribution is reflected in the rewards widget on the new-tab page (NTP).
- [ ] Verify monthly statement shows expected data.
- [ ] Verify when you click on the BR panel while on a site, the panel displays site-specific information (site favicon, domain, attention %).
- [ ] Verify BR panel shows message about an unverified publisher.
- [ ] Verify one time and monthly tip banners show a message about unverified publisher.
- [ ] Verify one time tip and monthly tip banners show a verified checkmark for a verified creator.
- [ ] Verify when you click on `Send a tip`, the custom tip banner displays if set up.
- [ ] Verify custom tip banner is also displayed for monthly contribution.
- [ ] Verify you are able to make one-time tip and they display in Tips panel.
- [ ] Verify tip is reflected in the rewards widget on the NTP.
- [ ] Verify when you tip an unverified publisher, the one time tip is recorded in the Pending Contributions list.
- [ ] Verify you are able to make recurring tip and they display in Monthly Contributions panel.
- [ ] Verify you are able to adjust your recurring tip amount from the BR panel.
- [ ] Verify recurring tip is reflected in the rewards widget on the NTP.
- [ ] Verify you can tip a verified website.
- [ ] Verify the website displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified YouTube creator.
- [ ] Verify the YouTube creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Vimeo creator.
- [ ] Verify the Vimeo creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Twitch creator.
- [ ] Verify the Twitch creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Twitter user from the panel.
- [ ] Verify you can tip a verified Twitter user via inline tip button.
- [ ] Verify the in-line tip button is spaced properly.
- [ ] Verify you can tip a verified GitHub user from the panel.
- [ ] Verify you can tip a verified GitHub user via inline tip button.
- [ ] Verify the GitHub creator displays in the auto-contribute list after specified amount of time/visits per settings.
- [ ] Verify you can tip a verified Reddit user from the panel.
- [ ] Verify you can tip a verified Reddit user via inline tip button.
- [ ] Verify if you disable auto-contribute you are still able to tip creators.
- [ ] Verify if auto-contribute is disabled AC does not occur.
- [ ] Verify if Rewards is disabled AC does not occur.
- [ ] Verify that disabling Rewards and enabling it again does not lose state.
- [ ] Adjust min visit/time in settings. Visit some sites to verify they are added to the table after the specified settings.
- [ ] Uphold cases
- [ ] Verify you are able to connect a KYC'd Uphold account to Rewards.
- [ ] Verify balance in Brave updates when BAT is added to the Brave Browser card.
- [ ] Verify if you only have user-controlled BAT (BAT in Uphold only), you can only tip KYC'd creators, any tips to non-KYC'd creators go to the Pending Contributions list.
- [ ] Verify connected (verified but not KYC'd) publishers display messaging on panel and tip banner.
- [ ] Verify you are able to perform an auto contribute using Uphold BAT.
### Social-media blocking settings
- [ ] Verify individual `Social media blocking` buttons works as intended when enabled/disabled by visiting https://fmarier.github.io/brave-testing/social-widgets.html
- [ ] visit `brave://settings/privacy` -> `Site and Shields Settings` -> `Cookies and site data` and ensure that
- [ ] both `https://[*.]firebaseapp.com` & `https://accounts.google.com` are added into `Sites that can always use cookies` when `Allow Google login buttons on third party sites` is enabled
- [ ] both `https://[*.]firebaseapp.com` & `https://accounts.google.com` are removed from `Sites that can always use cookies` when `Allow Google login buttons on third party sites` is disabled
- [ ] ensure that you can log in into https://www.expensify.com while `Allow Google login buttons on third party sites` is enabled
- [ ] ensure that once `Allow Google login buttons on third party sites` has been disabled, you can't log in into https://www.expensify.com
### Sync
- [ ] Verify you are able to create a sync chain and add a mobile/computer to the chain
- [ ] Verify you are able to join an existing sync chain using code words
- [ ] Verify the device name is shown properly when sync chain is created
- [ ] Verify you are able to add a new mobile device to the chain via QR code/code words
- [ ] Verify newly created bookmarks get sync'd to all devices on the sync chain
- [ ] Verify existing bookmarks on current profile gets sync'd to all devices on the sync chain
- [ ] Verify folder structure is retained after sync completes
- [ ] Verify bookmarks don't duplicate when sync'd from other devices
- [ ] Verify removing bookmark from device gets sync'd to all devices on the sync chain
- [ ] Verify adding/removing a bookmark in offline mode gets sync'd to all devices on the sync chain when device comes online
- [ ] With only two devices in chain, verify removing the other device resets the sync on b-c as well
### Tor Tabs
- [ ] Visit https://check.torproject.org in a Tor window, ensure it shows a success message for using a Tor exit node
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads
- [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Click `New Tor connection for this site` in app menu, ensure the exit node IP address changes after page is reloaded
- [ ] Visit https://protonmailrmez3lotccipshtkleegetolb73fuirgj7r4o4vfu7ozyd.onion, https://brave4u7jddbv7cyviptqjc7jusxh72uik7zt6adtckl5f4nwy2v72qd.onion/, and https://search.brave4u7jddbv7cyviptqjc7jusxh72uik7zt6adtckl5f4nwy2v72qd.onion/ in a Tor window and ensure all pages resolve
- [ ] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown
- [ ] Verify Torrent viewer doesn't load in a Tor window
- [ ] Ensure you are able to download a file in a Tor window. Verify all Download/Cancel, Download/Retry and Download works in Tor window
### Cookie and Cache
- [ ] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the evercookie site does not remember the old evercookie value
### Chromium/Brave GPU
- [ ] Verify that `brave://gpu` (Brave) matches `chrome://gpu` (Chrome) when using the same Chromium version
### Startup & Components
- [ ] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler`, `Wireshark` or `LittleSnitch` (or a similar application)
- [ ] Verify that opening a NTP doesn't trigger any outbound connections related to widgets without user interaction
- [ ] Remove the following component folders and ensure that they're being re-downloaded after restarting the browser:
- [ ] `afalakplffnnnlkncjhbmahjfjhmlkal`: `AutoplayWhitelist.dat`, `ExtensionWhitelist.dat`, `ReferrerWhitelist.json` and `Greaselion.json`
- [ ] `CertificateRevocation`
- [ ] `cffkpbalmllkdoenhmdmpbkajipdjfam`: `rs-ABPFilterParserData.dat` & `regional_catalog.json` (AdBlock)
- [ ] `gccbbckogglekeggclmmekihdgdpdgoe`: (Sponsored New Tab Images)
- [ ] `jicbkmdloagakknpihibphagfckhjdih`: `speedreader-updater.dat`
- [ ] `oofiananboodjbbmdelgdommihjbkfag`: HTTPSE
- [ ] `Safe Browsing`
- [ ] Restart the browser, load `brave://components`, wait for 8 mins and verify that no component shows any errors
**Note:** Always double check `brave://components` to make sure there's no errors/missing version numbers
### Session storage
- [ ] Temporarily move away your browser profile and test that a new profile is created on browser relaunch
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [ ] Test that both windows and tabs are being restored, including the current active tab
- [ ] Ensure that tabs are being lazy loaded when a previous session is being restored
### Upgrade
- [ ] Make sure that data from the last version appears in the new version OK
- [ ] Ensure that `brave://version` lists the expected Brave & Chromium versions
- [ ] With data from the last version, verify that:
- [ ] Bookmarks on the bookmark toolbar and bookmark folders can be opened
- [ ] Cookies are preserved
- [ ] Installed extensions are retained and work correctly
- [ ] Opened tabs can be reloaded
- [ ] Stored passwords are preserved
- [ ] Sync chain created in previous version is retained
- [ ] Social media-blocking buttons changes are retained
- [ ] Rewards
- [ ] BAT balance is retained
- [ ] Auto-contribute list is retained
- [ ] Both Tips and Monthly Contributions are retained
- [ ] Panel transactions list is retained
- [ ] Changes to rewards settings are retained
- [ ] Ensure that Auto Contribute is not being enabled when upgrading to a new version if AC was disabled
- [ ] Ads
- [ ] Both `Estimated pending rewards` & `Ad notifications received this month` are retained
- [ ] Changes to ads settings are retained
- [ ] Ensure that ads are not being enabled when upgrading to a new version if they were disabled
- [ ] Ensure that ads are not disabled when upgrading to a new version if they were enabled
## Hardware Wallet Test (To be checked on each major CR bump)
- [ ] Verify you can link hardware wallet using Trezor and unlock the wallet
- [ ] Verify you can link hardware wallet using Ledger Nano and unlock the wallet
- [ ] Verify you can perform a transaction with hardware wallet using Ledger Nano or Trezor
| test | manual test run on linux for x release installer check the installer is close to the size of the last release check signature if macos using binary run spctl assess verbose for the installed version and make sure it returns accepted if macos using universal binary run spctl assess verbose for the installed version and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window about pages verify that both chrome and about forward to brave run through several internal pages importing verify that you can import history favorites bookmarks and passwords from google chrome verify that you can import history favorites bookmarks passwords search engines and autofill from firefox verify that you can import favorites bookmarks from microsoft edge verify that importing bookmarks using bookmark html file retains the folder structure on a clean profile context menus verify you can block a page element using block element via selector context menu item verify selecting manage custom filters opens brave adblock in a ntp verify removing the rule from brave adblock reflects the change on the website after reload extensions plugins verify pdfium torrent viewer extensions are installed automatically on fresh profile and cannot be disabled they don t show up in brave extensions verify older version of an extension gets updated to new version via google server verify that magnet links and torrent files correctly open webtorrent and you re able to download the file s tip free torrents available via chrome web store cws verify that installing from cws displays the brave has not reviewed the extension warning via the add extension modal verify that installing from cws doesn t display the brave has not reviewed the extension warning via the add extension modal pdf test that you can print a pdf test that pdf is loaded over https at test that pdf is loaded over http at test that opens without issues widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine verify widevine notification is shown when you visit hbo max for the first time test that you can stream on hbo max on a fresh profile after installing widevine if macos run the above widevine tests for both and universal builds geolocation check that works and shows correct location check that works but doesn t require an accurate location crash reporting check that loading brave crash brave gpucrash causes the new tab to crash check that brave crashes lists the uploaded crash report id once the report has been submitted verify the crash id matches the report on backtrace using rxid equal bravery settings verify that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to verify that toggling trackers ads blocked works as expected visit verify that safe browsing via our proxy works for all the listed items visit and then turn on script blocking page should not load test that party storage results are blank at when party cookies are blocked and not blank when party cookies are unblocked test that shows up as grey not red no mixed content scripts are run in brave settings security choose a dns provider from the providers listed under use secure dns load and verify your isp s dns resolvers aren t detected and shown only your chosen doh provider should appear open a new private window with tor load and verify your isp s dns resolvers aren t detected and shown tls pinning visit and verify a pinning error is displayed visit and verify a pinning error is not displayed fingerprint tests test that doesn t leak ip address for each option under settings privacy and security webrtc ip handling policy brave ads verify when you enable rewards from panel or brave rewards ads are enabled by default verify ads ui panel settings etc shows when in a region with ads support verify ads ui panel settings etc does not show when in a region without ads support verify the ads panel does show the sorry ads are not yet available in your region message verify when the system language is english the browser language is french and you are in one of the supported regions ad notifications are still served to you verify you are served ad notifications when ads are enabled verify ad earnings are reflected in the rewards widget on the ntp verify when ads are toggled off there are no ad messages in the logs verify when rewards are toggled off but ads were not explicitly toggled off there are no ads logs recorded verify view click dismiss landed ad notifications show in confirmations json verify pages you browse to are being classified in the logs verify tokens are redeemed by viewing the logs you can use rewards debug true to shorten redemption time verify ad is not shown if a tab is playing media and is only shown after it stops playing rewards verify that none of the reward endpoints are being contacted when a user visits a media publisher youtube com reddit com twitter com github com and hasn t interacted with rewards verify that rewards brave com pcdn brave com grant rewards brave com or api rewards brave com are not being contacted verify you are able to create a new rewards profile verify you are able to restore an old rewards profile verify rewards balance shows correct bat and usd value verify actions taken claiming grant tipping auto contribute display in panel transactions list verify ac monthly budget shows correct bat and usd value verify you are able to exclude a publisher from the auto contribute table and popup list of sites verify you are able to exclude a publisher by using the toggle on the rewards panel verify you are able to perform an auto contribution verify auto contribution is reflected in the rewards widget on the new tab page ntp verify monthly statement shows expected data verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify br panel shows message about an unverified publisher verify one time and monthly tip banners show a message about unverified publisher verify one time tip and monthly tip banners show a verified checkmark for a verified creator verify when you click on send a tip the custom tip banner displays if set up verify custom tip banner is also displayed for monthly contribution verify you are able to make one time tip and they display in tips panel verify tip is reflected in the rewards widget on the ntp verify when you tip an unverified publisher the one time tip is recorded in the pending contributions list verify you are able to make recurring tip and they display in monthly contributions panel verify you are able to adjust your recurring tip amount from the br panel verify recurring tip is reflected in the rewards widget on the ntp verify you can tip a verified website verify the website displays in the auto contribute list after specified amount of time visits per settings verify you can tip a verified youtube creator verify the youtube creator displays in the auto contribute list after specified amount of time visits per settings verify you can tip a verified vimeo creator verify the vimeo creator displays in the auto contribute list after specified amount of time visits per settings verify you can tip a verified twitch creator verify the twitch creator displays in the auto contribute list after specified amount of time visits per settings verify you can tip a verified twitter user from the panel verify you can tip a verified twitter user via inline tip button verify the in line tip button is spaced properly verify you can tip a verified github user from the panel verify you can tip a verified github user via inline tip button verify the github creator displays in the auto contribute list after specified amount of time visits per settings verify you can tip a verified reddit user from the panel verify you can tip a verified reddit user via inline tip button verify if you disable auto contribute you are still able to tip creators verify if auto contribute is disabled ac does not occur verify if rewards is disabled ac does not occur verify that disabling rewards and enabling it again does not lose state adjust min visit time in settings visit some sites to verify they are added to the table after the specified settings uphold cases verify you are able to connect a kyc d uphold account to rewards verify balance in brave updates when bat is added to the brave browser card verify if you only have user controlled bat bat in uphold only you can only tip kyc d creators any tips to non kyc d creators go to the pending contributions list verify connected verified but not kyc d publishers display messaging on panel and tip banner verify you are able to perform an auto contribute using uphold bat social media blocking settings verify individual social media blocking buttons works as intended when enabled disabled by visiting visit brave settings privacy site and shields settings cookies and site data and ensure that both https firebaseapp com are added into sites that can always use cookies when allow google login buttons on third party sites is enabled both https firebaseapp com are removed from sites that can always use cookies when allow google login buttons on third party sites is disabled ensure that you can log in into while allow google login buttons on third party sites is enabled ensure that once allow google login buttons on third party sites has been disabled you can t log in into sync verify you are able to create a sync chain and add a mobile computer to the chain verify you are able to join an existing sync chain using code words verify the device name is shown properly when sync chain is created verify you are able to add a new mobile device to the chain via qr code code words verify newly created bookmarks get sync d to all devices on the sync chain verify existing bookmarks on current profile gets sync d to all devices on the sync chain verify folder structure is retained after sync completes verify bookmarks don t duplicate when sync d from other devices verify removing bookmark from device gets sync d to all devices on the sync chain verify adding removing a bookmark in offline mode gets sync d to all devices on the sync chain when device comes online with only two devices in chain verify removing the other device resets the sync on b c as well tor tabs visit in a tor window ensure it shows a success message for using a tor exit node visit in a tor window note down exit node ip address do a hard refresh ctrl shift r cmd shift r ensure exit ip changes after page reloads visit in a tor window note down exit node ip address click new tor connection for this site in app menu ensure the exit node ip address changes after page is reloaded visit and in a tor window and ensure all pages resolve visit in a tor window ensure location isn t shown verify torrent viewer doesn t load in a tor window ensure you are able to download a file in a tor window verify all download cancel download retry and download works in tor window cookie and cache go to and set an evercookie check that going to prefs clearing site data and cache and going back to the evercookie site does not remember the old evercookie value chromium brave gpu verify that brave gpu brave matches chrome gpu chrome when using the same chromium version startup components verify that brave is only contacting brave com endpoints on first launch using either charles proxy fiddler wireshark or littlesnitch or a similar application verify that opening a ntp doesn t trigger any outbound connections related to widgets without user interaction remove the following component folders and ensure that they re being re downloaded after restarting the browser afalakplffnnnlkncjhbmahjfjhmlkal autoplaywhitelist dat extensionwhitelist dat referrerwhitelist json and greaselion json certificaterevocation cffkpbalmllkdoenhmdmpbkajipdjfam rs abpfilterparserdata dat regional catalog json adblock gccbbckogglekeggclmmekihdgdpdgoe sponsored new tab images jicbkmdloagakknpihibphagfckhjdih speedreader updater dat oofiananboodjbbmdelgdommihjbkfag httpse safe browsing restart the browser load brave components wait for mins and verify that no component shows any errors note always double check brave components to make sure there s no errors missing version numbers session storage temporarily move away your browser profile and test that a new profile is created on browser relaunch macos library application support bravesoftware windows userprofile appdata local bravesoftware linux ubuntu config bravesoftware test that both windows and tabs are being restored including the current active tab ensure that tabs are being lazy loaded when a previous session is being restored upgrade make sure that data from the last version appears in the new version ok ensure that brave version lists the expected brave chromium versions with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved sync chain created in previous version is retained social media blocking buttons changes are retained rewards bat balance is retained auto contribute list is retained both tips and monthly contributions are retained panel transactions list is retained changes to rewards settings are retained ensure that auto contribute is not being enabled when upgrading to a new version if ac was disabled ads both estimated pending rewards ad notifications received this month are retained changes to ads settings are retained ensure that ads are not being enabled when upgrading to a new version if they were disabled ensure that ads are not disabled when upgrading to a new version if they were enabled hardware wallet test to be checked on each major cr bump verify you can link hardware wallet using trezor and unlock the wallet verify you can link hardware wallet using ledger nano and unlock the wallet verify you can perform a transaction with hardware wallet using ledger nano or trezor | 1 |
6,040 | 21,940,581,337 | IssuesEvent | 2022-05-23 17:39:12 | pharmaverse/admiral | https://api.github.com/repos/pharmaverse/admiral | closed | Create workflow to automatically create man files | automation | The workflow should be triggered whenever something is pushed to `devel` or `master`, run `devtools::document()` and commited any updated file in the `man` folder. | 1.0 | Create workflow to automatically create man files - The workflow should be triggered whenever something is pushed to `devel` or `master`, run `devtools::document()` and commited any updated file in the `man` folder. | non_test | create workflow to automatically create man files the workflow should be triggered whenever something is pushed to devel or master run devtools document and commited any updated file in the man folder | 0 |
618,208 | 19,429,339,994 | IssuesEvent | 2021-12-21 10:06:12 | kyma-incubator/reconciler | https://api.github.com/repos/kyma-incubator/reconciler | closed | Change detection: determine whether a resource can be patched, needs special patching or has to be replaced | enhancement area/reconciler Epic priority/critical prio1 | **Description**
The reconciler is per default [patching resources in Kubernetes](https://kubernetes.io/docs/tasks/manage-kubernetes-objects/update-api-object-kubectl-patch/) (merging the provided manifest with the resource-data available in K8s).
This merging / patching approach has same drawbacks (e.g. the applied merge strategy per resource-field is defined in K8s code and not configurable from outside) and cannot be fully controlled up-front by the reconciler.
To address this limitation, the reconciler has to make proper decisions whether a resource can
* be regular patched by using Kubernetes default patching strategy
* needs a specific patch strategy to update the resource (e.g. JSON patch, strategic patch with retainKeys constraint etc.)
* has to be replaced to ensure the resource is in the expected state
The existing [update-strategy resolver](https://github.com/kyma-incubator/reconciler/blob/main/pkg/reconciler/kubernetes/internal/updatestrategyresolver.go) could be extended or replaced for introducing an enhanced decision mechanism.
_Following rules have to be implemented:_
* [ ] TBD: please create sub-issue and describe the rule in detail + add link to issue here
* [ ] TBD: please create sub-issue and describe the rule in detail + add link to issue here
* [ ] ...
AC:
* rules can be defined which decide how a resource has to be updated (patch, strategic-patch or replace)
* rules are verified by integration tests (simulating situation in K8s and verifying patched resources in K8s)
* rules defined in table above are implemented
**Reasons**
Address drawbacks when using merge strategy for updating Kubernetes resources.
**Attachments**
<!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
| 1.0 | Change detection: determine whether a resource can be patched, needs special patching or has to be replaced - **Description**
The reconciler is per default [patching resources in Kubernetes](https://kubernetes.io/docs/tasks/manage-kubernetes-objects/update-api-object-kubectl-patch/) (merging the provided manifest with the resource-data available in K8s).
This merging / patching approach has same drawbacks (e.g. the applied merge strategy per resource-field is defined in K8s code and not configurable from outside) and cannot be fully controlled up-front by the reconciler.
To address this limitation, the reconciler has to make proper decisions whether a resource can
* be regular patched by using Kubernetes default patching strategy
* needs a specific patch strategy to update the resource (e.g. JSON patch, strategic patch with retainKeys constraint etc.)
* has to be replaced to ensure the resource is in the expected state
The existing [update-strategy resolver](https://github.com/kyma-incubator/reconciler/blob/main/pkg/reconciler/kubernetes/internal/updatestrategyresolver.go) could be extended or replaced for introducing an enhanced decision mechanism.
_Following rules have to be implemented:_
* [ ] TBD: please create sub-issue and describe the rule in detail + add link to issue here
* [ ] TBD: please create sub-issue and describe the rule in detail + add link to issue here
* [ ] ...
AC:
* rules can be defined which decide how a resource has to be updated (patch, strategic-patch or replace)
* rules are verified by integration tests (simulating situation in K8s and verifying patched resources in K8s)
* rules defined in table above are implemented
**Reasons**
Address drawbacks when using merge strategy for updating Kubernetes resources.
**Attachments**
<!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
| non_test | change detection determine whether a resource can be patched needs special patching or has to be replaced description the reconciler is per default merging the provided manifest with the resource data available in this merging patching approach has same drawbacks e g the applied merge strategy per resource field is defined in code and not configurable from outside and cannot be fully controlled up front by the reconciler to address this limitation the reconciler has to make proper decisions whether a resource can be regular patched by using kubernetes default patching strategy needs a specific patch strategy to update the resource e g json patch strategic patch with retainkeys constraint etc has to be replaced to ensure the resource is in the expected state the existing could be extended or replaced for introducing an enhanced decision mechanism following rules have to be implemented tbd please create sub issue and describe the rule in detail add link to issue here tbd please create sub issue and describe the rule in detail add link to issue here ac rules can be defined which decide how a resource has to be updated patch strategic patch or replace rules are verified by integration tests simulating situation in and verifying patched resources in rules defined in table above are implemented reasons address drawbacks when using merge strategy for updating kubernetes resources attachments | 0 |
264,901 | 23,145,072,395 | IssuesEvent | 2022-07-28 23:13:23 | MPMG-DCC-UFMG/F01 | https://api.github.com/repos/MPMG-DCC-UFMG/F01 | closed | Teste de generalizacao para a tag Seridores - Registro por lotação - Janaúba | generalization test development template-Síntese tecnologia informatica tag-Servidores subtag-Registro por lotação | DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Janaúba. | 1.0 | Teste de generalizacao para a tag Seridores - Registro por lotação - Janaúba - DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Janaúba. | test | teste de generalizacao para a tag seridores registro por lotação janaúba dod realizar o teste de generalização do validador da tag seridores registro por lotação para o município de janaúba | 1 |
270,965 | 23,550,451,018 | IssuesEvent | 2022-08-21 18:56:41 | 4ian/GDevelop | https://api.github.com/repos/4ian/GDevelop | closed | Crash while using an editor | Need a game/precise steps to reproduce the issue 👋 Needs confirmation/testing wontfix | I'll be honest, I have no clue what happened. I was messing around in the piskel editor when this happened. Piskel was occasionally freezing, meaning I'd have to save my animation every now and then just in case.
## Error stack (don't write anything here)
```
RuntimeError: memory access out of bounds
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/libGD.wasm?cache-buster=5.0.138-81ef11163db977d9e3821c16568930651e50c580:wasm-function[92]:0x206cc
at Resource.GetName.Resource.GetName [as getName] (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/libGD.js?cache-buster=5.0.138-81ef11163db977d9e3821c16568930651e50c580:9:860295)
at _ (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:1217524)
at n.value (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:161413)
at children (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:163007)
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:81292
at Yo (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:394056)
at za (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:402948)
at ks (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:441398)
at mu (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:432922)
at vu (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:432847)
at su (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:429856)
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381334
at t.unstable_runWithPriority (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:456561)
at $i (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381043)
at Qi (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381279)
at Ji (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381214)
at nu (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:426638)
at ba (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:398429)
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/main.d0c0ad0b.chunk.js:1:19922
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/main.d0c0ad0b.chunk.js:1:97975
at Set.forEach (<anonymous>)
at s (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/main.d0c0ad0b.chunk.js:1:97947)
at i (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:568873)
```
## Component stack (don't write anything here)
```
in Unknown
in DropTarget(Component)
in DragSource(DropTarget(Component))
in Unknown
in div
in div
in div
in t
in t
in o
in c
in n
in t
in div
in t
in div
in div
in ForwardRef
in ForwardRef
in Unknown
in n
in div
in div
in t
in div
in t
in Unknown
in n
in div
in n
in n
in t
in Unknown
in Unknown
in div
in n
in Unknown
in div
in Unknown
in n
in d
in j
in b
in b
in h
in g
in c
in n
in t
in pe
in n
in Unknown
in g
in t
in l
in o
in n
in n
in DragDropContext(n)
in n
in c
```
## Other details
* IDE version: 5.0.138-81ef11163db977d9e3821c16568930651e50c580
* Arch: x64,
* Platform Name: win32,
* System Version: 10.0.19043,
* User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) GDevelop5/5.0.138 Chrome/100.0.4896.143 Electron/18.2.2 Safari/537.36,
| 1.0 | Crash while using an editor - I'll be honest, I have no clue what happened. I was messing around in the piskel editor when this happened. Piskel was occasionally freezing, meaning I'd have to save my animation every now and then just in case.
## Error stack (don't write anything here)
```
RuntimeError: memory access out of bounds
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/libGD.wasm?cache-buster=5.0.138-81ef11163db977d9e3821c16568930651e50c580:wasm-function[92]:0x206cc
at Resource.GetName.Resource.GetName [as getName] (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/libGD.js?cache-buster=5.0.138-81ef11163db977d9e3821c16568930651e50c580:9:860295)
at _ (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:1217524)
at n.value (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:161413)
at children (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:163007)
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/1.f636414a.chunk.js:1:81292
at Yo (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:394056)
at za (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:402948)
at ks (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:441398)
at mu (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:432922)
at vu (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:432847)
at su (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:429856)
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381334
at t.unstable_runWithPriority (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:456561)
at $i (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381043)
at Qi (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381279)
at Ji (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:381214)
at nu (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:426638)
at ba (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:398429)
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/main.d0c0ad0b.chunk.js:1:19922
at file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/main.d0c0ad0b.chunk.js:1:97975
at Set.forEach (<anonymous>)
at s (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/main.d0c0ad0b.chunk.js:1:97947)
at i (file:///C:/Users/User%201/AppData/Local/Programs/GDevelop/resources/app.asar/www/static/js/64.a0030cea.chunk.js:2:568873)
```
## Component stack (don't write anything here)
```
in Unknown
in DropTarget(Component)
in DragSource(DropTarget(Component))
in Unknown
in div
in div
in div
in t
in t
in o
in c
in n
in t
in div
in t
in div
in div
in ForwardRef
in ForwardRef
in Unknown
in n
in div
in div
in t
in div
in t
in Unknown
in n
in div
in n
in n
in t
in Unknown
in Unknown
in div
in n
in Unknown
in div
in Unknown
in n
in d
in j
in b
in b
in h
in g
in c
in n
in t
in pe
in n
in Unknown
in g
in t
in l
in o
in n
in n
in DragDropContext(n)
in n
in c
```
## Other details
* IDE version: 5.0.138-81ef11163db977d9e3821c16568930651e50c580
* Arch: x64,
* Platform Name: win32,
* System Version: 10.0.19043,
* User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) GDevelop5/5.0.138 Chrome/100.0.4896.143 Electron/18.2.2 Safari/537.36,
| test | crash while using an editor i ll be honest i have no clue what happened i was messing around in the piskel editor when this happened piskel was occasionally freezing meaning i d have to save my animation every now and then just in case error stack don t write anything here runtimeerror memory access out of bounds at file c users user appdata local programs gdevelop resources app asar www libgd wasm cache buster wasm function at resource getname resource getname file c users user appdata local programs gdevelop resources app asar www libgd js cache buster at file c users user appdata local programs gdevelop resources app asar www static js chunk js at n value file c users user appdata local programs gdevelop resources app asar www static js chunk js at children file c users user appdata local programs gdevelop resources app asar www static js chunk js at file c users user appdata local programs gdevelop resources app asar www static js chunk js at yo file c users user appdata local programs gdevelop resources app asar www static js chunk js at za file c users user appdata local programs gdevelop resources app asar www static js chunk js at ks file c users user appdata local programs gdevelop resources app asar www static js chunk js at mu file c users user appdata local programs gdevelop resources app asar www static js chunk js at vu file c users user appdata local programs gdevelop resources app asar www static js chunk js at su file c users user appdata local programs gdevelop resources app asar www static js chunk js at file c users user appdata local programs gdevelop resources app asar www static js chunk js at t unstable runwithpriority file c users user appdata local programs gdevelop resources app asar www static js chunk js at i file c users user appdata local programs gdevelop resources app asar www static js chunk js at qi file c users user appdata local programs gdevelop resources app asar www static js chunk js at ji file c users user appdata local programs gdevelop resources app asar www static js chunk js at nu file c users user appdata local programs gdevelop resources app asar www static js chunk js at ba file c users user appdata local programs gdevelop resources app asar www static js chunk js at file c users user appdata local programs gdevelop resources app asar www static js main chunk js at file c users user appdata local programs gdevelop resources app asar www static js main chunk js at set foreach at s file c users user appdata local programs gdevelop resources app asar www static js main chunk js at i file c users user appdata local programs gdevelop resources app asar www static js chunk js component stack don t write anything here in unknown in droptarget component in dragsource droptarget component in unknown in div in div in div in t in t in o in c in n in t in div in t in div in div in forwardref in forwardref in unknown in n in div in div in t in div in t in unknown in n in div in n in n in t in unknown in unknown in div in n in unknown in div in unknown in n in d in j in b in b in h in g in c in n in t in pe in n in unknown in g in t in l in o in n in n in dragdropcontext n in n in c other details ide version arch platform name system version user agent mozilla windows nt applewebkit khtml like gecko chrome electron safari | 1 |
60,905 | 3,135,575,583 | IssuesEvent | 2015-09-10 15:49:12 | kesava-wow/kuinameplates | https://api.github.com/repos/kesava-wow/kuinameplates | opened | Do the SECRET THING | enhancement low priority | By "secret thing", I mean my idea for an interface which would make it easier to configure the positions of frame elements with drag-dropping, etc.
Secret. | 1.0 | Do the SECRET THING - By "secret thing", I mean my idea for an interface which would make it easier to configure the positions of frame elements with drag-dropping, etc.
Secret. | non_test | do the secret thing by secret thing i mean my idea for an interface which would make it easier to configure the positions of frame elements with drag dropping etc secret | 0 |
218,147 | 16,959,727,756 | IssuesEvent | 2021-06-29 00:47:38 | prestodb/presto | https://api.github.com/repos/prestodb/presto | closed | Tests failing due to float precision | available-to-pickup stale tests | On master (13e1d0bc30), `mvn clean test -pl :presto-geospatial` is failing for me on three tests. Two seem to be due to float precision:
```
[ERROR] Failures:
[ERROR] TestGeoFunctions.testGreatCircleDistance:819->AbstractTestFunctions.assertFunction:96 expected [2886.448973436703] but found [2886.4489734367016]
[ERROR] TestSphericalGeoFunctions.testDistance:139->assertDistance:151->AbstractTestFunctions.assertFunction:96 expected [2886448.973436703] but found [2886448.9734367016]
```
They are tested with `assertFunction`, which ultimately calls `assertEqual`, which is generally a bad choice for floats. Do we have a similar function with something like `assertAlmostEqual(..., precision)`?
cc: @tdcmeehan | 1.0 | Tests failing due to float precision - On master (13e1d0bc30), `mvn clean test -pl :presto-geospatial` is failing for me on three tests. Two seem to be due to float precision:
```
[ERROR] Failures:
[ERROR] TestGeoFunctions.testGreatCircleDistance:819->AbstractTestFunctions.assertFunction:96 expected [2886.448973436703] but found [2886.4489734367016]
[ERROR] TestSphericalGeoFunctions.testDistance:139->assertDistance:151->AbstractTestFunctions.assertFunction:96 expected [2886448.973436703] but found [2886448.9734367016]
```
They are tested with `assertFunction`, which ultimately calls `assertEqual`, which is generally a bad choice for floats. Do we have a similar function with something like `assertAlmostEqual(..., precision)`?
cc: @tdcmeehan | test | tests failing due to float precision on master mvn clean test pl presto geospatial is failing for me on three tests two seem to be due to float precision failures testgeofunctions testgreatcircledistance abstracttestfunctions assertfunction expected but found testsphericalgeofunctions testdistance assertdistance abstracttestfunctions assertfunction expected but found they are tested with assertfunction which ultimately calls assertequal which is generally a bad choice for floats do we have a similar function with something like assertalmostequal precision cc tdcmeehan | 1 |
240,307 | 20,022,087,482 | IssuesEvent | 2022-02-01 17:17:46 | vector-im/matrix-video-chat | https://api.github.com/repos/vector-im/matrix-video-chat | closed | Failed camera / microphone / screenshare permission requests | T-Bug p1 Z-Community-Testing | From Kat:
> After failing to join on desktop because it doesn't ask for camera permissions (in Firefox), I can't join on Android for the same reason
> "Firefox did not ask for permission to use camera after some testing (nothing else was using the camera at the time)
I think even through I was on the call preview screen, it was still sending my audio to the call? Not sure"
> Can't join call if I don't have a camera
> Can't join call if I don't have a mic | 1.0 | Failed camera / microphone / screenshare permission requests - From Kat:
> After failing to join on desktop because it doesn't ask for camera permissions (in Firefox), I can't join on Android for the same reason
> "Firefox did not ask for permission to use camera after some testing (nothing else was using the camera at the time)
I think even through I was on the call preview screen, it was still sending my audio to the call? Not sure"
> Can't join call if I don't have a camera
> Can't join call if I don't have a mic | test | failed camera microphone screenshare permission requests from kat after failing to join on desktop because it doesn t ask for camera permissions in firefox i can t join on android for the same reason firefox did not ask for permission to use camera after some testing nothing else was using the camera at the time i think even through i was on the call preview screen it was still sending my audio to the call not sure can t join call if i don t have a camera can t join call if i don t have a mic | 1 |
384,384 | 11,388,311,857 | IssuesEvent | 2020-01-29 16:28:01 | netdata/netdata | https://api.github.com/repos/netdata/netdata | opened | Properly handle DM multipath on Linux in `system-info.sh` when collecting disk sizes. | area/daemon bug priority/medium | Currently, we completely ignore device-mapper devices on Linux when calculating disk sizes using the contents of `/sys/block`. This is usually a good thing, but it means that we overreport disk sizes on systems that use multipathing.
For at least the case of DM multipath (the common case), we can work around this by parsing the DM table for each DM target to see if it's a multipath target, and if so use the target device and ignore the component devices. | 1.0 | Properly handle DM multipath on Linux in `system-info.sh` when collecting disk sizes. - Currently, we completely ignore device-mapper devices on Linux when calculating disk sizes using the contents of `/sys/block`. This is usually a good thing, but it means that we overreport disk sizes on systems that use multipathing.
For at least the case of DM multipath (the common case), we can work around this by parsing the DM table for each DM target to see if it's a multipath target, and if so use the target device and ignore the component devices. | non_test | properly handle dm multipath on linux in system info sh when collecting disk sizes currently we completely ignore device mapper devices on linux when calculating disk sizes using the contents of sys block this is usually a good thing but it means that we overreport disk sizes on systems that use multipathing for at least the case of dm multipath the common case we can work around this by parsing the dm table for each dm target to see if it s a multipath target and if so use the target device and ignore the component devices | 0 |
214,626 | 16,601,588,925 | IssuesEvent | 2021-06-01 20:16:58 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | Port Project management functionality from Ember UI | QA/M [zube]: To Test tmp-smart | Create equivalent functionality in the Dashboard UI to allow Projects to be managed.
This should continue to use the Norman API.
This currently looks like this:

Broadly we can keep the same look and feel - use the same tab style for the project view as used in the namespace view - e.g.

We need to support the actions on namespaces in projects to allow them to be moved to a different project.
There is quite a bit of overlap between the project view and the namespace view, so we should consider this when implementing this feature.
| 1.0 | Port Project management functionality from Ember UI - Create equivalent functionality in the Dashboard UI to allow Projects to be managed.
This should continue to use the Norman API.
This currently looks like this:

Broadly we can keep the same look and feel - use the same tab style for the project view as used in the namespace view - e.g.

We need to support the actions on namespaces in projects to allow them to be moved to a different project.
There is quite a bit of overlap between the project view and the namespace view, so we should consider this when implementing this feature.
| test | port project management functionality from ember ui create equivalent functionality in the dashboard ui to allow projects to be managed this should continue to use the norman api this currently looks like this broadly we can keep the same look and feel use the same tab style for the project view as used in the namespace view e g we need to support the actions on namespaces in projects to allow them to be moved to a different project there is quite a bit of overlap between the project view and the namespace view so we should consider this when implementing this feature | 1 |
82,431 | 15,646,560,769 | IssuesEvent | 2021-03-23 01:12:39 | jgeraigery/linux | https://api.github.com/repos/jgeraigery/linux | opened | CVE-2020-14390 (Medium) detected in linuxv5.2 | security vulnerability | ## CVE-2020-14390 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/drivers/video/fbdev/core/fbcon.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/drivers/video/fbdev/core/fbcon.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel in versions before 5.9-rc6. When changing screen size, an out-of-bounds memory write can occur leading to memory corruption or a denial of service. Due to the nature of the flaw, privilege escalation cannot be fully ruled out.
<p>Publish Date: 2020-09-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14390>CVE-2020-14390</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14390">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14390</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: v5.9-rc6,v5.8.10,v5.4.66</p>
</p>
</details>
<p></p>
| True | CVE-2020-14390 (Medium) detected in linuxv5.2 - ## CVE-2020-14390 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/drivers/video/fbdev/core/fbcon.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/drivers/video/fbdev/core/fbcon.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel in versions before 5.9-rc6. When changing screen size, an out-of-bounds memory write can occur leading to memory corruption or a denial of service. Due to the nature of the flaw, privilege escalation cannot be fully ruled out.
<p>Publish Date: 2020-09-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14390>CVE-2020-14390</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14390">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14390</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: v5.9-rc6,v5.8.10,v5.4.66</p>
</p>
</details>
<p></p>
| non_test | cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href vulnerable source files linux drivers video fbdev core fbcon c linux drivers video fbdev core fbcon c vulnerability details a flaw was found in the linux kernel in versions before when changing screen size an out of bounds memory write can occur leading to memory corruption or a denial of service due to the nature of the flaw privilege escalation cannot be fully ruled out publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
324,863 | 27,826,139,550 | IssuesEvent | 2023-03-19 19:41:26 | ikeman2003/Dreams-Of-Peaches-new-video-automation | https://api.github.com/repos/ikeman2003/Dreams-Of-Peaches-new-video-automation | closed | The event named "test" occurred on the Maker Webhooks service | test | <tr class="public-audio">
<td>
<br>
<a href="https://youtu.be/wUK9ZzRkmfo" target="_blank">Best Friend Is Jealous of You Dating</a><br>
</td>
<td>[Friends to lovers] [Confession] [Emotional] [Wholesome]</td>
</tr>
<br> | 1.0 | The event named "test" occurred on the Maker Webhooks service - <tr class="public-audio">
<td>
<br>
<a href="https://youtu.be/wUK9ZzRkmfo" target="_blank">Best Friend Is Jealous of You Dating</a><br>
</td>
<td>[Friends to lovers] [Confession] [Emotional] [Wholesome]</td>
</tr>
<br> | test | the event named test occurred on the maker webhooks service best friend is jealous of you dating | 1 |
269,870 | 23,472,988,432 | IssuesEvent | 2022-08-17 01:02:47 | CDL-Dryad/dryad-product-roadmap | https://api.github.com/repos/CDL-Dryad/dryad-product-roadmap | closed | Investigate related works: absent and duplicates | Testing | In one case the related works will not appear for Zenodo and in another there are duplicates showing for SI and software. This only happened when all three types of things are submitted. | 1.0 | Investigate related works: absent and duplicates - In one case the related works will not appear for Zenodo and in another there are duplicates showing for SI and software. This only happened when all three types of things are submitted. | test | investigate related works absent and duplicates in one case the related works will not appear for zenodo and in another there are duplicates showing for si and software this only happened when all three types of things are submitted | 1 |
236,968 | 7,754,299,118 | IssuesEvent | 2018-05-31 06:00:57 | kubeflow/tf-operator | https://api.github.com/repos/kubeflow/tf-operator | closed | TfJob operator stops working on invalid spec | api/v1alpha1 api/v1alpha2 area/operator kind/bug priority/p1 | I submitted a job with an invalid spec (container args contained integrs and not strings). The job was created but it was never started and the status was never updated. Furthermore, I think this blocked the TFJob operator from processing any other jobs. Deleting the job fixed things.
The TFJob operator showed the following logs.
```
E0422 01:21:04.083809 1 reflector.go:205] github.com/kubeflow/tf-operator/pkg/client/informers/externalversions/factory.go:59: Failed to list *v1alpha1.TFJob: v1alpha1.TFJobList: Items: []v1alpha1.TFJob: v1alpha1.TFJob: Spec: v1alpha1.TFJobSpec: ReplicaSpecs: []*v1alpha1.TFReplicaSpec: v1alpha1.TFReplicaSpec: Template: v1.PodTemplateSpec: Spec: v1.PodSpec: Containers: []v1.Container: v1.Container: Env: []v1.EnvVar: v1.EnvVar: v1.EnvVar: Value: ReadString: expects " or n, parsing 1996 ...,"value":1... at {"apiVersion":"kubeflow.org/v1alpha1","items":[{"apiVersion":"kubeflow.org/v1alpha1","kind":"TF
```
I believe what's happening is that since the spec is invalid the result of List can't be successfully parsed into a Go struct. As a result, I think the TFJob operator is unable to work.
I think this is a problem in the underlying informer package; i.e. its not robust to invalid specs. We should check if this is a known issue and if there is an existing bug. (Ideally, it would just ignore invalid specs).
I think we could fix this a number of ways in TFJob controller
- If we use CRD's spec validation feature and provide a swagger spec I think we could prevent invalid specs from being accepted in the first place
- The operator could try to catch this error and then find and update the invalid spec
- Deleting the job might be confusing for users but maybe we could leave the job there but delete most fields and set status to "invalid spec"?
Swagger is probably the best place to start.
We should try to get this fixed in 0.2 | 1.0 | TfJob operator stops working on invalid spec - I submitted a job with an invalid spec (container args contained integrs and not strings). The job was created but it was never started and the status was never updated. Furthermore, I think this blocked the TFJob operator from processing any other jobs. Deleting the job fixed things.
The TFJob operator showed the following logs.
```
E0422 01:21:04.083809 1 reflector.go:205] github.com/kubeflow/tf-operator/pkg/client/informers/externalversions/factory.go:59: Failed to list *v1alpha1.TFJob: v1alpha1.TFJobList: Items: []v1alpha1.TFJob: v1alpha1.TFJob: Spec: v1alpha1.TFJobSpec: ReplicaSpecs: []*v1alpha1.TFReplicaSpec: v1alpha1.TFReplicaSpec: Template: v1.PodTemplateSpec: Spec: v1.PodSpec: Containers: []v1.Container: v1.Container: Env: []v1.EnvVar: v1.EnvVar: v1.EnvVar: Value: ReadString: expects " or n, parsing 1996 ...,"value":1... at {"apiVersion":"kubeflow.org/v1alpha1","items":[{"apiVersion":"kubeflow.org/v1alpha1","kind":"TF
```
I believe what's happening is that since the spec is invalid the result of List can't be successfully parsed into a Go struct. As a result, I think the TFJob operator is unable to work.
I think this is a problem in the underlying informer package; i.e. its not robust to invalid specs. We should check if this is a known issue and if there is an existing bug. (Ideally, it would just ignore invalid specs).
I think we could fix this a number of ways in TFJob controller
- If we use CRD's spec validation feature and provide a swagger spec I think we could prevent invalid specs from being accepted in the first place
- The operator could try to catch this error and then find and update the invalid spec
- Deleting the job might be confusing for users but maybe we could leave the job there but delete most fields and set status to "invalid spec"?
Swagger is probably the best place to start.
We should try to get this fixed in 0.2 | non_test | tfjob operator stops working on invalid spec i submitted a job with an invalid spec container args contained integrs and not strings the job was created but it was never started and the status was never updated furthermore i think this blocked the tfjob operator from processing any other jobs deleting the job fixed things the tfjob operator showed the following logs reflector go github com kubeflow tf operator pkg client informers externalversions factory go failed to list tfjob tfjoblist items tfjob tfjob spec tfjobspec replicaspecs tfreplicaspec tfreplicaspec template podtemplatespec spec podspec containers container container env envvar envvar envvar value readstring expects or n parsing value at apiversion kubeflow org items apiversion kubeflow org kind tf i believe what s happening is that since the spec is invalid the result of list can t be successfully parsed into a go struct as a result i think the tfjob operator is unable to work i think this is a problem in the underlying informer package i e its not robust to invalid specs we should check if this is a known issue and if there is an existing bug ideally it would just ignore invalid specs i think we could fix this a number of ways in tfjob controller if we use crd s spec validation feature and provide a swagger spec i think we could prevent invalid specs from being accepted in the first place the operator could try to catch this error and then find and update the invalid spec deleting the job might be confusing for users but maybe we could leave the job there but delete most fields and set status to invalid spec swagger is probably the best place to start we should try to get this fixed in | 0 |
331,477 | 28,965,019,610 | IssuesEvent | 2023-05-10 07:15:39 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix container.test_container_structural_diff | Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_misc/test_container.py::test_container_structural_diff[cpu-ivy.functional.backends.tensorflow-False-False]</summary>
2023-05-10T05:03:55.1297006Z E AttributeError: 'Shape' object has no attribute '_dims'2023-05-10T05:03:55.1305152Z E ivy.utils.exceptions.IvyError: tensorflow: all_equal: 'Shape' object has no attribute '_dims'
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_misc/test_container.py::test_container_structural_diff[cpu-ivy.functional.backends.tensorflow-False-False]</summary>
2023-05-10T05:03:55.1297006Z E AttributeError: 'Shape' object has no attribute '_dims'2023-05-10T05:03:55.1305152Z E ivy.utils.exceptions.IvyError: tensorflow: all_equal: 'Shape' object has no attribute '_dims'
</details>
| 1.0 | Fix container.test_container_structural_diff - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4933154251/jobs/8816856131" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_misc/test_container.py::test_container_structural_diff[cpu-ivy.functional.backends.tensorflow-False-False]</summary>
2023-05-10T05:03:55.1297006Z E AttributeError: 'Shape' object has no attribute '_dims'2023-05-10T05:03:55.1305152Z E ivy.utils.exceptions.IvyError: tensorflow: all_equal: 'Shape' object has no attribute '_dims'
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_misc/test_container.py::test_container_structural_diff[cpu-ivy.functional.backends.tensorflow-False-False]</summary>
2023-05-10T05:03:55.1297006Z E AttributeError: 'Shape' object has no attribute '_dims'2023-05-10T05:03:55.1305152Z E ivy.utils.exceptions.IvyError: tensorflow: all_equal: 'Shape' object has no attribute '_dims'
</details>
| test | fix container test container structural diff tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test misc test container py test container structural diff e attributeerror shape object has no attribute dims e ivy utils exceptions ivyerror tensorflow all equal shape object has no attribute dims failed ivy tests test ivy test misc test container py test container structural diff e attributeerror shape object has no attribute dims e ivy utils exceptions ivyerror tensorflow all equal shape object has no attribute dims | 1 |
165,344 | 26,150,301,725 | IssuesEvent | 2022-12-30 12:33:33 | YAPP-Github/21st-Web-Team-1-FE | https://api.github.com/repos/YAPP-Github/21st-Web-Team-1-FE | closed | ✨ 메인 페이지 및 검색 페이지 UI 작업 | feature design | ### 정보
- 담당자 : @SeieunYoo
- 개요 :
- 기한 :
### 설명
mock api 로 인기 태그 구현 및 메인 페이지 검색 페이지 UI 작업 마무리
이번에는 작업 볼륨 작게 가져갑니다~.


이번에는 작업 볼륨 작게 가져갑니다~.
### 체크리스트
- [x] mock api 인기 태그
- [x] 메인 페이지 UI
- [x] 검색 페이지 UI
- [ ] <추후 고려 사항> SEO | 1.0 | ✨ 메인 페이지 및 검색 페이지 UI 작업 - ### 정보
- 담당자 : @SeieunYoo
- 개요 :
- 기한 :
### 설명
mock api 로 인기 태그 구현 및 메인 페이지 검색 페이지 UI 작업 마무리
이번에는 작업 볼륨 작게 가져갑니다~.


이번에는 작업 볼륨 작게 가져갑니다~.
### 체크리스트
- [x] mock api 인기 태그
- [x] 메인 페이지 UI
- [x] 검색 페이지 UI
- [ ] <추후 고려 사항> SEO | non_test | ✨ 메인 페이지 및 검색 페이지 ui 작업 정보 담당자 seieunyoo 개요 기한 설명 mock api 로 인기 태그 구현 및 메인 페이지 검색 페이지 ui 작업 마무리 이번에는 작업 볼륨 작게 가져갑니다 이번에는 작업 볼륨 작게 가져갑니다 체크리스트 mock api 인기 태그 메인 페이지 ui 검색 페이지 ui seo | 0 |
59,724 | 6,661,909,011 | IssuesEvent | 2017-10-02 10:45:06 | fabric8-services/fabric8-wit | https://api.github.com/repos/fabric8-services/fabric8-wit | opened | Random test failures because of work item creation error | database test | Sometimes, a test fails with the following unique constraint error:
````
pq: duplicate key value violates unique constraint "work_item_number_sequences_pkey"
````
This seems to happen when a couple of work items are created in a new space (eg: https://github.com/fabric8-services/fabric8-wit/blob/master/search/search_repository_whitebox_test.go#L53)
| 1.0 | Random test failures because of work item creation error - Sometimes, a test fails with the following unique constraint error:
````
pq: duplicate key value violates unique constraint "work_item_number_sequences_pkey"
````
This seems to happen when a couple of work items are created in a new space (eg: https://github.com/fabric8-services/fabric8-wit/blob/master/search/search_repository_whitebox_test.go#L53)
| test | random test failures because of work item creation error sometimes a test fails with the following unique constraint error pq duplicate key value violates unique constraint work item number sequences pkey this seems to happen when a couple of work items are created in a new space eg | 1 |
122,660 | 10,229,003,632 | IssuesEvent | 2019-08-17 08:31:44 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | system:readDir lists files and folder in sub directories in windows. | Area/StandardLibs BetaTesting Priority/High Type/Bug | **Description:**
When using the `system:readDir` to get the directories of a folder it gives files and folders of subdirectories as well. Ideally it should only gives files and folders of the given path.
It also outputs the current folder as well.
Suggestion: May be we can pass in another arg stating recursive or not.
**Steps to reproduce:**
```ballerina
system:FileInfo[] files = check system:readDir(path);
foreach system:FileInfo file in files {
string isDirectory = file.isDir() ? "true" : "false";
log:printInfo(file.getName() + " - " + isDirectory);
}
```
**Affected Versions:**
**OS, DB, other environment details and versions:**
Windows | 1.0 | system:readDir lists files and folder in sub directories in windows. - **Description:**
When using the `system:readDir` to get the directories of a folder it gives files and folders of subdirectories as well. Ideally it should only gives files and folders of the given path.
It also outputs the current folder as well.
Suggestion: May be we can pass in another arg stating recursive or not.
**Steps to reproduce:**
```ballerina
system:FileInfo[] files = check system:readDir(path);
foreach system:FileInfo file in files {
string isDirectory = file.isDir() ? "true" : "false";
log:printInfo(file.getName() + " - " + isDirectory);
}
```
**Affected Versions:**
**OS, DB, other environment details and versions:**
Windows | test | system readdir lists files and folder in sub directories in windows description when using the system readdir to get the directories of a folder it gives files and folders of subdirectories as well ideally it should only gives files and folders of the given path it also outputs the current folder as well suggestion may be we can pass in another arg stating recursive or not steps to reproduce ballerina system fileinfo files check system readdir path foreach system fileinfo file in files string isdirectory file isdir true false log printinfo file getname isdirectory affected versions os db other environment details and versions windows | 1 |
341,524 | 30,590,784,486 | IssuesEvent | 2023-07-21 16:51:16 | denoland/fresh | https://api.github.com/repos/denoland/fresh | closed | Make windows CI tests more reliable | kind:tests | Currently, the windows CI is failing frequently. This looks to be a timing issue as re-running it usually makes it green again. With https://github.com/denoland/fresh/pull/1473 I tried to make the tests less sensitive to timing issues, but it's there needs to be more work done as the windows CI keeps failing often. | 1.0 | Make windows CI tests more reliable - Currently, the windows CI is failing frequently. This looks to be a timing issue as re-running it usually makes it green again. With https://github.com/denoland/fresh/pull/1473 I tried to make the tests less sensitive to timing issues, but it's there needs to be more work done as the windows CI keeps failing often. | test | make windows ci tests more reliable currently the windows ci is failing frequently this looks to be a timing issue as re running it usually makes it green again with i tried to make the tests less sensitive to timing issues but it s there needs to be more work done as the windows ci keeps failing often | 1 |
373,015 | 26,031,789,538 | IssuesEvent | 2022-12-21 22:11:09 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | Add TOC to help find flags in Commmand Line Reference | type: documentation (cleanup) P2 team-Documentation | Documentation URL: https://docs.bazel.build/versions/master/command-line-reference.html
Very hard to tell what command a flag applies to at a glance. If you take a flag in the middle of the page, you have to scroll way up to find which section it's in and make sure not to miss it.
Is it possible to add a TOC at the very top? | 2.0 | Add TOC to help find flags in Commmand Line Reference - Documentation URL: https://docs.bazel.build/versions/master/command-line-reference.html
Very hard to tell what command a flag applies to at a glance. If you take a flag in the middle of the page, you have to scroll way up to find which section it's in and make sure not to miss it.
Is it possible to add a TOC at the very top? | non_test | add toc to help find flags in commmand line reference documentation url very hard to tell what command a flag applies to at a glance if you take a flag in the middle of the page you have to scroll way up to find which section it s in and make sure not to miss it is it possible to add a toc at the very top | 0 |
78,757 | 7,663,903,485 | IssuesEvent | 2018-05-12 00:10:40 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | opened | [CI] SmokeTestMonitoringWithSecurityIT fails due to leaked thread | :Security/Network >test-failure | SmokeTestMonitoringWithSecurityIT failed due to leaked threads.
```
2> SEVERE: 1 thread leaked from SUITE scope at org.elasticsearch.smoketest.SmokeTestMonitoringWithSecurityIT:
2> 1) Thread[id=18, name=threadDeathWatcher-2-1, state=TIMED_WAITING, group=TGRP-SmokeTestMonitoringWithSecurityIT]
2> at java.lang.Thread.sleep(Native Method)
2> at io.netty.util.ThreadDeathWatcher$Watcher.run(ThreadDeathWatcher.java:152)
2> at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)
2> at java.lang.Thread.run(Thread.java:748)
```
- Instance: https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob-windows-compatibility/1627/console
- Log: [testHTTPExporterWithSSL.txt](https://github.com/elastic/elasticsearch/files/1997121/testHTTPExporterWithSSL.txt) | 1.0 | [CI] SmokeTestMonitoringWithSecurityIT fails due to leaked thread - SmokeTestMonitoringWithSecurityIT failed due to leaked threads.
```
2> SEVERE: 1 thread leaked from SUITE scope at org.elasticsearch.smoketest.SmokeTestMonitoringWithSecurityIT:
2> 1) Thread[id=18, name=threadDeathWatcher-2-1, state=TIMED_WAITING, group=TGRP-SmokeTestMonitoringWithSecurityIT]
2> at java.lang.Thread.sleep(Native Method)
2> at io.netty.util.ThreadDeathWatcher$Watcher.run(ThreadDeathWatcher.java:152)
2> at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)
2> at java.lang.Thread.run(Thread.java:748)
```
- Instance: https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob-windows-compatibility/1627/console
- Log: [testHTTPExporterWithSSL.txt](https://github.com/elastic/elasticsearch/files/1997121/testHTTPExporterWithSSL.txt) | test | smoketestmonitoringwithsecurityit fails due to leaked thread smoketestmonitoringwithsecurityit failed due to leaked threads severe thread leaked from suite scope at org elasticsearch smoketest smoketestmonitoringwithsecurityit thread at java lang thread sleep native method at io netty util threaddeathwatcher watcher run threaddeathwatcher java at io netty util concurrent defaultthreadfactory defaultrunnabledecorator run defaultthreadfactory java at java lang thread run thread java instance log | 1 |
204,989 | 15,577,190,829 | IssuesEvent | 2021-03-17 13:16:09 | tarantool/metrics | https://api.github.com/repos/tarantool/metrics | closed | Broken Promtool test | bug ci tests | Promtool test is broken:
- https://github.com/prometheus/prometheus/issues/5590 - may be fixed as it mentioned in ticket
- troubles with tarantool installer (tarantool seems not to be installed) - may be fixed by Tarantool-install action | 1.0 | Broken Promtool test - Promtool test is broken:
- https://github.com/prometheus/prometheus/issues/5590 - may be fixed as it mentioned in ticket
- troubles with tarantool installer (tarantool seems not to be installed) - may be fixed by Tarantool-install action | test | broken promtool test promtool test is broken may be fixed as it mentioned in ticket troubles with tarantool installer tarantool seems not to be installed may be fixed by tarantool install action | 1 |
117,877 | 11,958,095,677 | IssuesEvent | 2020-04-04 16:47:27 | glotzerlab/signac-flow | https://api.github.com/repos/glotzerlab/signac-flow | closed | Add doc linting to CI | documentation enhancement | <!-- Please replace the text in the individual sections below. -->
### Feature description
Replicate glotzerlab/signac#294 for this repo. Pick some file as a starting point (we used jsondict.py in signac) and go from there.
### Proposed solution
Use pydocstyle as described in the corresponding signac issue. | 1.0 | Add doc linting to CI - <!-- Please replace the text in the individual sections below. -->
### Feature description
Replicate glotzerlab/signac#294 for this repo. Pick some file as a starting point (we used jsondict.py in signac) and go from there.
### Proposed solution
Use pydocstyle as described in the corresponding signac issue. | non_test | add doc linting to ci feature description replicate glotzerlab signac for this repo pick some file as a starting point we used jsondict py in signac and go from there proposed solution use pydocstyle as described in the corresponding signac issue | 0 |
62,815 | 6,817,622,375 | IssuesEvent | 2017-11-07 00:21:30 | sybrew/the-seo-framework | https://api.github.com/repos/sybrew/the-seo-framework | opened | Incorrect description pixel counter description | Testing | In the image below you can see a truncated description:

You can also see timestamps. That's great, but I think they add to the pixel limit. Ergo the truncation.
I've tested this in the pixel counter by adding the timestamp manually. It might be that there's reserved space (e.g. the length of `Dec 10, 2017` instead of just `1 day ago`).
This means we should lower the description pixel limit value from `920` to `838` on posts.
But, they may remain `920` on pages.

I need to test this further... | 1.0 | Incorrect description pixel counter description - In the image below you can see a truncated description:

You can also see timestamps. That's great, but I think they add to the pixel limit. Ergo the truncation.
I've tested this in the pixel counter by adding the timestamp manually. It might be that there's reserved space (e.g. the length of `Dec 10, 2017` instead of just `1 day ago`).
This means we should lower the description pixel limit value from `920` to `838` on posts.
But, they may remain `920` on pages.

I need to test this further... | test | incorrect description pixel counter description in the image below you can see a truncated description you can also see timestamps that s great but i think they add to the pixel limit ergo the truncation i ve tested this in the pixel counter by adding the timestamp manually it might be that there s reserved space e g the length of dec instead of just day ago this means we should lower the description pixel limit value from to on posts but they may remain on pages i need to test this further | 1 |
95,368 | 10,878,813,983 | IssuesEvent | 2019-11-16 20:19:36 | mpostol/OPC-UA-OOI | https://api.github.com/repos/mpostol/OPC-UA-OOI | closed | Add reference to `Object Oriented Internet` conference paper. | Documentation In progress In waiting | Add a reference to the conference paper:
[**Object Oriented Internet**](https://annals-csis.org/proceedings/2015/drp/160.html), Proceedings of the 2015 Federated Conference on Computer Science and Information Systems
DOI: 10.15439/2015F160,
2015,
| 1.0 | Add reference to `Object Oriented Internet` conference paper. - Add a reference to the conference paper:
[**Object Oriented Internet**](https://annals-csis.org/proceedings/2015/drp/160.html), Proceedings of the 2015 Federated Conference on Computer Science and Information Systems
DOI: 10.15439/2015F160,
2015,
| non_test | add reference to object oriented internet conference paper add a reference to the conference paper proceedings of the federated conference on computer science and information systems doi | 0 |
196,413 | 14,858,730,427 | IssuesEvent | 2021-01-18 17:14:50 | mindsdb/mindsdb | https://api.github.com/repos/mindsdb/mindsdb | closed | Timeseries-focused prediction latency test | Clickhouse timeseries tests | Create a test script inside the `tests` directory, which we can run manually and that does the following two things:
1. Trains models from 2x timeseries datasets (@paxcema please recommend the two datasets to be used :d) ... unless the models already exist, in which case no need to re-train every time the script runs
2. For both models run a bunch of predict queries via mindsdb native's `predict` function:
* 1 by 1 predictions using data in a clickhouse table (i.e. using a `ClickhouseDS` datasource)
* bulk predictions using data in a clickhouse table (i.e. using a `ClickhouseDS` datasource) ... bulk meaning selecting, say, between 20 and 100 rows
* 1 by 1 predictions using data passed in the form of a dictionary/pandas dataframe/csv file, including some historical data
* bulk preidctions using data passed in the form p a pandas dataframe or csv file ... again, bulk here predicting for something like 20 to 100 rows.
3. Run the above 4 test cases (or at least the first 2) using the mysql proxy via a clickhouse integration (i.e. sending `SELECT` queries from within clickhouse to the predictor tables created within the mindsdb database)
4. The same thing as 3 and 2, but this time using the python sdk
When this is all done, generate a report detailing how long each of the tasks took in order to compare the latency between these 2 interfaces and mindsdb_native itself, as well as to benchmark the time native needs to run timeseries predictions.
| 1.0 | Timeseries-focused prediction latency test - Create a test script inside the `tests` directory, which we can run manually and that does the following two things:
1. Trains models from 2x timeseries datasets (@paxcema please recommend the two datasets to be used :d) ... unless the models already exist, in which case no need to re-train every time the script runs
2. For both models run a bunch of predict queries via mindsdb native's `predict` function:
* 1 by 1 predictions using data in a clickhouse table (i.e. using a `ClickhouseDS` datasource)
* bulk predictions using data in a clickhouse table (i.e. using a `ClickhouseDS` datasource) ... bulk meaning selecting, say, between 20 and 100 rows
* 1 by 1 predictions using data passed in the form of a dictionary/pandas dataframe/csv file, including some historical data
* bulk preidctions using data passed in the form p a pandas dataframe or csv file ... again, bulk here predicting for something like 20 to 100 rows.
3. Run the above 4 test cases (or at least the first 2) using the mysql proxy via a clickhouse integration (i.e. sending `SELECT` queries from within clickhouse to the predictor tables created within the mindsdb database)
4. The same thing as 3 and 2, but this time using the python sdk
When this is all done, generate a report detailing how long each of the tasks took in order to compare the latency between these 2 interfaces and mindsdb_native itself, as well as to benchmark the time native needs to run timeseries predictions.
| test | timeseries focused prediction latency test create a test script inside the tests directory which we can run manually and that does the following two things trains models from timeseries datasets paxcema please recommend the two datasets to be used d unless the models already exist in which case no need to re train every time the script runs for both models run a bunch of predict queries via mindsdb native s predict function by predictions using data in a clickhouse table i e using a clickhouseds datasource bulk predictions using data in a clickhouse table i e using a clickhouseds datasource bulk meaning selecting say between and rows by predictions using data passed in the form of a dictionary pandas dataframe csv file including some historical data bulk preidctions using data passed in the form p a pandas dataframe or csv file again bulk here predicting for something like to rows run the above test cases or at least the first using the mysql proxy via a clickhouse integration i e sending select queries from within clickhouse to the predictor tables created within the mindsdb database the same thing as and but this time using the python sdk when this is all done generate a report detailing how long each of the tasks took in order to compare the latency between these interfaces and mindsdb native itself as well as to benchmark the time native needs to run timeseries predictions | 1 |
333,219 | 29,516,579,327 | IssuesEvent | 2023-06-04 15:00:42 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix statistical.test_sum | Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix statistical.test_sum - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5163430059/jobs/9301746563" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
| test | fix statistical test sum tensorflow img src torch img src numpy img src jax img src paddle img src | 1 |
208,676 | 15,914,248,867 | IssuesEvent | 2021-04-13 00:11:34 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Flaky test UnexpectedRequestFrame | HTTP3 area-runtime test-failure | `Microsoft.AspNetCore.Server.Kestrel.Core.Tests.Http3StreamTests.UnexpectedRequestFrame(frameType: "CancelPush")`
> Assert.Equal() Failure
Expected: UnexpectedFrame
Actual: 0
> at Microsoft.AspNetCore.Server.Kestrel.Core.Tests.Http3TestBase.Http3RequestStream.WaitForStreamErrorAsync(Http3ErrorCode protocolError, String expectedErrorMessage) in /_/src/Servers/Kestrel/test/InMemory.FunctionalTests/Http3/Http3TestBase.cs:line 458
at Microsoft.AspNetCore.Server.Kestrel.Core.Tests.Http3StreamTests.UnexpectedRequestFrame(String frameType) in /_/src/Servers/Kestrel/test/InMemory.FunctionalTests/Http3/Http3StreamTests.cs:line 1759
--- End of stack trace from previous location ---
Failed here: https://dev.azure.com/dnceng/public/_build/results?buildId=1063023&view=results | 1.0 | Flaky test UnexpectedRequestFrame - `Microsoft.AspNetCore.Server.Kestrel.Core.Tests.Http3StreamTests.UnexpectedRequestFrame(frameType: "CancelPush")`
> Assert.Equal() Failure
Expected: UnexpectedFrame
Actual: 0
> at Microsoft.AspNetCore.Server.Kestrel.Core.Tests.Http3TestBase.Http3RequestStream.WaitForStreamErrorAsync(Http3ErrorCode protocolError, String expectedErrorMessage) in /_/src/Servers/Kestrel/test/InMemory.FunctionalTests/Http3/Http3TestBase.cs:line 458
at Microsoft.AspNetCore.Server.Kestrel.Core.Tests.Http3StreamTests.UnexpectedRequestFrame(String frameType) in /_/src/Servers/Kestrel/test/InMemory.FunctionalTests/Http3/Http3StreamTests.cs:line 1759
--- End of stack trace from previous location ---
Failed here: https://dev.azure.com/dnceng/public/_build/results?buildId=1063023&view=results | test | flaky test unexpectedrequestframe microsoft aspnetcore server kestrel core tests unexpectedrequestframe frametype cancelpush assert equal failure expected unexpectedframe actual at microsoft aspnetcore server kestrel core tests waitforstreamerrorasync protocolerror string expectederrormessage in src servers kestrel test inmemory functionaltests cs line at microsoft aspnetcore server kestrel core tests unexpectedrequestframe string frametype in src servers kestrel test inmemory functionaltests cs line end of stack trace from previous location failed here | 1 |
320,864 | 27,489,750,513 | IssuesEvent | 2023-03-04 13:20:48 | Nasanin-Parast/Rollenspiel | https://api.github.com/repos/Nasanin-Parast/Rollenspiel | closed | Spieler implementieren | Sprint 1 Sprint 2 Implementierung Test | **Aufgabenstellung**
Der in Issue #15 erstellte Entwurf ist zu implementieren und zu testen.
**Inhalt**
wird durch Issue #15 noch genauer definiert -
**Akzeptanzkriterien**
- [x] Der Spieler ist implementiert
- [x] Der Spieler ist durch Unit-Tests abgedeckt
- [x] Die Diagramme zum Spieler sind angepasst
- [x] Die Änderungen sind erfolgreich nach master gebracht
- [x] Dokumentation ist dazu geschrieben
**Verknüpfte Issues**
#15
**Offene Fragen** | 1.0 | Spieler implementieren - **Aufgabenstellung**
Der in Issue #15 erstellte Entwurf ist zu implementieren und zu testen.
**Inhalt**
wird durch Issue #15 noch genauer definiert -
**Akzeptanzkriterien**
- [x] Der Spieler ist implementiert
- [x] Der Spieler ist durch Unit-Tests abgedeckt
- [x] Die Diagramme zum Spieler sind angepasst
- [x] Die Änderungen sind erfolgreich nach master gebracht
- [x] Dokumentation ist dazu geschrieben
**Verknüpfte Issues**
#15
**Offene Fragen** | test | spieler implementieren aufgabenstellung der in issue erstellte entwurf ist zu implementieren und zu testen inhalt wird durch issue noch genauer definiert akzeptanzkriterien der spieler ist implementiert der spieler ist durch unit tests abgedeckt die diagramme zum spieler sind angepasst die änderungen sind erfolgreich nach master gebracht dokumentation ist dazu geschrieben verknüpfte issues offene fragen | 1 |
31,079 | 4,682,630,006 | IssuesEvent | 2016-10-09 11:00:27 | leelokas/vacation_tracker | https://api.github.com/repos/leelokas/vacation_tracker | closed | Vacation overview view | IN TESTING P2 Task Vacation list | There has to be a page visible to all users that contains info about all the vacations that are either in “PLANNED” or “CONFIRMED” stages | 1.0 | Vacation overview view - There has to be a page visible to all users that contains info about all the vacations that are either in “PLANNED” or “CONFIRMED” stages | test | vacation overview view there has to be a page visible to all users that contains info about all the vacations that are either in “planned” or “confirmed” stages | 1 |
107,721 | 9,221,746,806 | IssuesEvent | 2019-03-11 20:47:21 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: X-Pack Spaces API Integration Tests -- spaces_only.x-pack/test/spaces_api_integration/spaces_only/apis/get_all·ts - spaces api without security get all can access all spaces from default "before all" hook | failed-test | A test failed on a tracked branch
```
Error: Unable to fetch Kibana status API response from Kibana at http://elastic:changeme@localhost:5620: Error: Client request error: connect ECONNREFUSED 127.0.0.1:5620
at getKibanaPluginEnabled (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup5/node/immutable/kibana/src/es_archiver/lib/indices/kibana_index.js:138:11)
at process._tickCallback (internal/process/next_tick.js:68:7)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=x-pack-ciGroup5,node=immutable/409/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Spaces API Integration Tests -- spaces_only.x-pack/test/spaces_api_integration/spaces_only/apis/get_all·ts","test.name":"spaces api without security get all can access all spaces from default \"before all\" hook","test.failCount":1}} --> | 1.0 | Failing test: X-Pack Spaces API Integration Tests -- spaces_only.x-pack/test/spaces_api_integration/spaces_only/apis/get_all·ts - spaces api without security get all can access all spaces from default "before all" hook - A test failed on a tracked branch
```
Error: Unable to fetch Kibana status API response from Kibana at http://elastic:changeme@localhost:5620: Error: Client request error: connect ECONNREFUSED 127.0.0.1:5620
at getKibanaPluginEnabled (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup5/node/immutable/kibana/src/es_archiver/lib/indices/kibana_index.js:138:11)
at process._tickCallback (internal/process/next_tick.js:68:7)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=x-pack-ciGroup5,node=immutable/409/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Spaces API Integration Tests -- spaces_only.x-pack/test/spaces_api_integration/spaces_only/apis/get_all·ts","test.name":"spaces api without security get all can access all spaces from default \"before all\" hook","test.failCount":1}} --> | test | failing test x pack spaces api integration tests spaces only x pack test spaces api integration spaces only apis get all·ts spaces api without security get all can access all spaces from default before all hook a test failed on a tracked branch error unable to fetch kibana status api response from kibana at error client request error connect econnrefused at getkibanapluginenabled var lib jenkins workspace elastic kibana master job x pack node immutable kibana src es archiver lib indices kibana index js at process tickcallback internal process next tick js first failure | 1 |
36,083 | 5,030,466,360 | IssuesEvent | 2016-12-16 00:56:31 | EpochModTeam/Epoch | https://api.github.com/repos/EpochModTeam/Epoch | closed | AH F5 delete nomore works | bug Feedback Needed Needs Tested | F5 nomore kicks you out of the game, but delete function disappears after half a second, so I can't delete anything anymore.. | 1.0 | AH F5 delete nomore works - F5 nomore kicks you out of the game, but delete function disappears after half a second, so I can't delete anything anymore.. | test | ah delete nomore works nomore kicks you out of the game but delete function disappears after half a second so i can t delete anything anymore | 1 |
77,529 | 7,575,624,279 | IssuesEvent | 2018-04-24 02:45:30 | awyand/Tandm | https://api.github.com/repos/awyand/Tandm | closed | add button to add preset missions, phones, and containers to a user | task testing | Hard code a few example missions (with phones, with containers) on click for testing purposes to make sure database is set up properly. | 1.0 | add button to add preset missions, phones, and containers to a user - Hard code a few example missions (with phones, with containers) on click for testing purposes to make sure database is set up properly. | test | add button to add preset missions phones and containers to a user hard code a few example missions with phones with containers on click for testing purposes to make sure database is set up properly | 1 |
63,201 | 14,656,678,321 | IssuesEvent | 2020-12-28 13:57:42 | fu1771695yongxie/strapi | https://api.github.com/repos/fu1771695yongxie/strapi | opened | CVE-2020-11023 (Medium) detected in jquery-1.7.2.min.js | security vulnerability | ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: strapi/packages/strapi-provider-upload-rackspace/node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: strapi/packages/strapi-provider-upload-rackspace/node_modules/jmespath/index.html,strapi/packages/strapi-provider-upload-aws-s3/node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/strapi/commit/ef946669e47731733c9539896862ef97eceeb459">ef946669e47731733c9539896862ef97eceeb459</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-11023 (Medium) detected in jquery-1.7.2.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: strapi/packages/strapi-provider-upload-rackspace/node_modules/jmespath/index.html</p>
<p>Path to vulnerable library: strapi/packages/strapi-provider-upload-rackspace/node_modules/jmespath/index.html,strapi/packages/strapi-provider-upload-aws-s3/node_modules/jmespath/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/strapi/commit/ef946669e47731733c9539896862ef97eceeb459">ef946669e47731733c9539896862ef97eceeb459</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file strapi packages strapi provider upload rackspace node modules jmespath index html path to vulnerable library strapi packages strapi provider upload rackspace node modules jmespath index html strapi packages strapi provider upload aws node modules jmespath index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
622,025 | 19,604,599,035 | IssuesEvent | 2022-01-06 07:43:13 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | olmoauth.outlook.com - site is not usable | browser-chrome priority-normal QA_triaged | <!-- @browser: Chrome 96.0.4664 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/96.0.4664.110 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/97809 -->
**URL**: https://olmoauth.outlook.com
**Browser / Version**: Chrome 96.0.4664
**Operating System**: Mac OS X 10.15.7
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
i'm just adding new account on outlook, but at the last part of adding, the website has been crashed
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/4b928b6e-40b8-474f-872f-0c095c0c6684.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | olmoauth.outlook.com - site is not usable - <!-- @browser: Chrome 96.0.4664 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/96.0.4664.110 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/97809 -->
**URL**: https://olmoauth.outlook.com
**Browser / Version**: Chrome 96.0.4664
**Operating System**: Mac OS X 10.15.7
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
i'm just adding new account on outlook, but at the last part of adding, the website has been crashed
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/4b928b6e-40b8-474f-872f-0c095c0c6684.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | olmoauth outlook com site is not usable url browser version chrome operating system mac os x tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce i m just adding new account on outlook but at the last part of adding the website has been crashed view the screenshot img alt screenshot src browser configuration none from with ❤️ | 0 |
121,212 | 10,153,528,678 | IssuesEvent | 2019-08-06 05:00:20 | wijehyeon/Easy-way-2-Use-FFMPEG | https://api.github.com/repos/wijehyeon/Easy-way-2-Use-FFMPEG | closed | VideoTransformService 테스트 추가 | Test | * [x] : makeFileByLocalTime 테스트 코드
* [x] : videoTransformCommandGenerator 테스트 코드
* [x] : mp4ToGif 테스트 코드 | 1.0 | VideoTransformService 테스트 추가 - * [x] : makeFileByLocalTime 테스트 코드
* [x] : videoTransformCommandGenerator 테스트 코드
* [x] : mp4ToGif 테스트 코드 | test | videotransformservice 테스트 추가 makefilebylocaltime 테스트 코드 videotransformcommandgenerator 테스트 코드 테스트 코드 | 1 |
90,353 | 8,234,197,637 | IssuesEvent | 2018-09-08 11:29:42 | imixs/imixs-workflow | https://api.github.com/repos/imixs/imixs-workflow | closed | XMLDataCollectionAdapter - check for null values | bug testing | method getDataCollection should check for null values | 1.0 | XMLDataCollectionAdapter - check for null values - method getDataCollection should check for null values | test | xmldatacollectionadapter check for null values method getdatacollection should check for null values | 1 |
47,581 | 5,903,520,527 | IssuesEvent | 2017-05-19 07:03:59 | EenmaalAndermaal/EenmaalAndermaal | https://api.github.com/repos/EenmaalAndermaal/EenmaalAndermaal | closed | Functioneel ontwerp | tester: Bart tester: Jesper tester: Sven tester: Thijs tester: Wouter tester: Yuri | # Gerelateerde issues
- #
- #
# Omschrijving
Wensen en eisen van de veilingwebsite.
Usecases en activity diagrams. | 6.0 | Functioneel ontwerp - # Gerelateerde issues
- #
- #
# Omschrijving
Wensen en eisen van de veilingwebsite.
Usecases en activity diagrams. | test | functioneel ontwerp gerelateerde issues omschrijving wensen en eisen van de veilingwebsite usecases en activity diagrams | 1 |
92,470 | 15,857,083,320 | IssuesEvent | 2021-04-08 03:55:33 | heholek/better-onetab | https://api.github.com/repos/heholek/better-onetab | opened | CVE-2020-24025 (Medium) detected in node-sass-4.14.1.tgz | security vulnerability | ## CVE-2020-24025 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: better-onetab/package.json</p>
<p>Path to vulnerable library: better-onetab/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Certificate validation in node-sass 2.0.0 to 4.14.1 is disabled when requesting binaries even if the user is not specifying an alternative download path.
<p>Publish Date: 2021-01-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24025>CVE-2020-24025</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-24025 (Medium) detected in node-sass-4.14.1.tgz - ## CVE-2020-24025 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: better-onetab/package.json</p>
<p>Path to vulnerable library: better-onetab/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Certificate validation in node-sass 2.0.0 to 4.14.1 is disabled when requesting binaries even if the user is not specifying an alternative download path.
<p>Publish Date: 2021-01-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24025>CVE-2020-24025</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in node sass tgz cve medium severity vulnerability vulnerable library node sass tgz wrapper around libsass library home page a href path to dependency file better onetab package json path to vulnerable library better onetab node modules node sass package json dependency hierarchy x node sass tgz vulnerable library vulnerability details certificate validation in node sass to is disabled when requesting binaries even if the user is not specifying an alternative download path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href step up your open source security game with whitesource | 0 |
123,716 | 10,280,454,049 | IssuesEvent | 2019-08-26 05:24:21 | kubernetes/test-infra | https://api.github.com/repos/kubernetes/test-infra | closed | Lacking in location config when creating cluster during e2e tests. | kind/bug sig/cloud-provider sig/testing | <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**What happened**:
When creating a cluster during an e2e test, there's an error report showing the `location` of the resource group is missing.
**What you expected to happen**:
The cluster should be created successfully.
**How to reproduce it (as minimally and precisely as possible)**:
Enable any azure e2e test.
**Please provide links to example occurrences, if any**:
**Anything else we need to know?**:
/sig cloud-provider
/sig testing
| 1.0 | Lacking in location config when creating cluster during e2e tests. - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**What happened**:
When creating a cluster during an e2e test, there's an error report showing the `location` of the resource group is missing.
**What you expected to happen**:
The cluster should be created successfully.
**How to reproduce it (as minimally and precisely as possible)**:
Enable any azure e2e test.
**Please provide links to example occurrences, if any**:
**Anything else we need to know?**:
/sig cloud-provider
/sig testing
| test | lacking in location config when creating cluster during tests what happened when creating a cluster during an test there s an error report showing the location of the resource group is missing what you expected to happen the cluster should be created successfully how to reproduce it as minimally and precisely as possible enable any azure test please provide links to example occurrences if any anything else we need to know sig cloud provider sig testing | 1 |
30,042 | 4,546,845,508 | IssuesEvent | 2016-09-12 00:54:17 | backdrop/backdrop-issues | https://api.github.com/repos/backdrop/backdrop-issues | closed | [UX] Provide a Hero block feature out of the box. | pr - reviewed & tested by the community status - has pull request type - feature request | Alright, I know that this can be achieved by contrib modules like [Nivo Slider](https://github.com/backdrop-contrib/nivo_slider) or [Views Slideshow](https://github.com/backdrop-contrib/views_slideshow), but I believe it is very common nowadays for people to want to design their website with either a single image section (banner) or a slider.
The single image (banner) section can be seen as the simplest version of the slider that only contains one slide. So, would it be too far-fetched to ask that this feature is offered by Backdrop out of the box?
This was kinda brought up in #1042 and I think would be a killer feature for Backdrop!! ...but because I see this easily ending in an endless debate over which slideshow module/library is "the best", I propose to base this on a view (already in core) of nodes that contain only image and description/link fields (soon in core) and make the slider library "plugable".
---
~~PR from @jenlampton https://github.com/backdrop/backdrop/issues/1523~~
~~PR from @quicksketch https://github.com/backdrop/backdrop/issues/1526~~
PR from @jenlampton https://github.com/backdrop/backdrop/pull/1528 | 1.0 | [UX] Provide a Hero block feature out of the box. - Alright, I know that this can be achieved by contrib modules like [Nivo Slider](https://github.com/backdrop-contrib/nivo_slider) or [Views Slideshow](https://github.com/backdrop-contrib/views_slideshow), but I believe it is very common nowadays for people to want to design their website with either a single image section (banner) or a slider.
The single image (banner) section can be seen as the simplest version of the slider that only contains one slide. So, would it be too far-fetched to ask that this feature is offered by Backdrop out of the box?
This was kinda brought up in #1042 and I think would be a killer feature for Backdrop!! ...but because I see this easily ending in an endless debate over which slideshow module/library is "the best", I propose to base this on a view (already in core) of nodes that contain only image and description/link fields (soon in core) and make the slider library "plugable".
---
~~PR from @jenlampton https://github.com/backdrop/backdrop/issues/1523~~
~~PR from @quicksketch https://github.com/backdrop/backdrop/issues/1526~~
PR from @jenlampton https://github.com/backdrop/backdrop/pull/1528 | test | provide a hero block feature out of the box alright i know that this can be achieved by contrib modules like or but i believe it is very common nowadays for people to want to design their website with either a single image section banner or a slider the single image banner section can be seen as the simplest version of the slider that only contains one slide so would it be too far fetched to ask that this feature is offered by backdrop out of the box this was kinda brought up in and i think would be a killer feature for backdrop but because i see this easily ending in an endless debate over which slideshow module library is the best i propose to base this on a view already in core of nodes that contain only image and description link fields soon in core and make the slider library plugable pr from jenlampton pr from quicksketch pr from jenlampton | 1 |
159,860 | 12,494,444,638 | IssuesEvent | 2020-06-01 11:13:04 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: acceptance/bank/cluster-recovery failed | C-test-failure O-roachtest O-robot branch-master release-blocker | [(roachtest).acceptance/bank/cluster-recovery failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1982079&tab=buildLog) on [master@187036c55297a0ded5703c7b969c05736e11e363](https://github.com/cockroachdb/cockroach/commits/187036c55297a0ded5703c7b969c05736e11e363):
```
| | * Consider using 'cockroach start-single-node' or 'cockroach init' instead.
| | *
| | *
| | * ERROR: ERROR: connection lost.
| | *
| | * cockroach server exited with error: failed to create engines: unexpected EOF
| | *
| | ERROR: connection lost.
| |
| | cockroach server exited with error: failed to create engines: unexpected EOF
| | Failed running "start"
| | E200601 11:07:10.900642 1 cli/error.go:374 ERROR: exit status 1
| | ERROR: exit status 1
| | Failed running "start"
| Wraps: (4) exit status 1
| Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *exec.ExitError:
| I200601 11:07:10.905426 1 cluster_synced.go:1749 command failed
|
| stdout:
| local: starting
Wraps: (2) exit status 1
Error types: (1) *main.withCommandDetails (2) *exec.ExitError
bank.go:371,bank.go:448,acceptance.go:90,test_runner.go:753: context canceled
cluster.go:1512,context.go:135,cluster.go:1501,test_runner.go:825: dead node detection: /go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor local --oneshot --ignore-empty-nodes: exit status 1 3: 506
2: 785
1: dead
4: 882
Error: UNCLASSIFIED_PROBLEM: 1: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
| main.glob..func13
| /go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1129
| main.wrap.func1
| /go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:272
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute
| /go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC
| /go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute
| /go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800
| main.main
| /go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1799
| runtime.main
| /usr/local/go/src/runtime/proc.go:203
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (3) 1: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/acceptance/bank/cluster-recovery](https://teamcity.cockroachdb.com/viewLog.html?buildId=1982079&tab=artifacts#/acceptance/bank/cluster-recovery)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aacceptance%2Fbank%2Fcluster-recovery.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 2.0 | roachtest: acceptance/bank/cluster-recovery failed - [(roachtest).acceptance/bank/cluster-recovery failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1982079&tab=buildLog) on [master@187036c55297a0ded5703c7b969c05736e11e363](https://github.com/cockroachdb/cockroach/commits/187036c55297a0ded5703c7b969c05736e11e363):
```
| | * Consider using 'cockroach start-single-node' or 'cockroach init' instead.
| | *
| | *
| | * ERROR: ERROR: connection lost.
| | *
| | * cockroach server exited with error: failed to create engines: unexpected EOF
| | *
| | ERROR: connection lost.
| |
| | cockroach server exited with error: failed to create engines: unexpected EOF
| | Failed running "start"
| | E200601 11:07:10.900642 1 cli/error.go:374 ERROR: exit status 1
| | ERROR: exit status 1
| | Failed running "start"
| Wraps: (4) exit status 1
| Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *exec.ExitError:
| I200601 11:07:10.905426 1 cluster_synced.go:1749 command failed
|
| stdout:
| local: starting
Wraps: (2) exit status 1
Error types: (1) *main.withCommandDetails (2) *exec.ExitError
bank.go:371,bank.go:448,acceptance.go:90,test_runner.go:753: context canceled
cluster.go:1512,context.go:135,cluster.go:1501,test_runner.go:825: dead node detection: /go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor local --oneshot --ignore-empty-nodes: exit status 1 3: 506
2: 785
1: dead
4: 882
Error: UNCLASSIFIED_PROBLEM: 1: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
| main.glob..func13
| /go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1129
| main.wrap.func1
| /go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:272
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute
| /go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC
| /go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute
| /go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800
| main.main
| /go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1799
| runtime.main
| /usr/local/go/src/runtime/proc.go:203
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (3) 1: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/acceptance/bank/cluster-recovery](https://teamcity.cockroachdb.com/viewLog.html?buildId=1982079&tab=artifacts#/acceptance/bank/cluster-recovery)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aacceptance%2Fbank%2Fcluster-recovery.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| test | roachtest acceptance bank cluster recovery failed on consider using cockroach start single node or cockroach init instead error error connection lost cockroach server exited with error failed to create engines unexpected eof error connection lost cockroach server exited with error failed to create engines unexpected eof failed running start cli error go error exit status error exit status failed running start wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage exec exiterror cluster synced go command failed stdout local starting wraps exit status error types main withcommanddetails exec exiterror bank go bank go acceptance go test runner go context canceled cluster go context go cluster go test runner go dead node detection go src github com cockroachdb cockroach bin roachprod monitor local oneshot ignore empty nodes exit status dead error unclassified problem dead unclassified problem wraps attached stack trace main glob go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap go src github com cockroachdb cockroach pkg cmd roachprod main go github com cockroachdb cockroach vendor github com cobra command execute go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command executec go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command execute go src github com cockroachdb cockroach vendor github com cobra command go main main go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types errors unclassified withstack withstack errors errorstring more artifacts powered by | 1 |
141,148 | 18,949,106,992 | IssuesEvent | 2021-11-18 13:29:14 | paulius-valiunas/dotenv-expand | https://api.github.com/repos/paulius-valiunas/dotenv-expand | opened | WS-2020-0042 (High) detected in acorn-5.7.3.tgz | security vulnerability | ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.3.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p>
<p>Path to dependency file: dotenv-expand/package.json</p>
<p>Path to vulnerable library: dotenv-expand/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- lab-13.1.0.tgz (Root Library)
- espree-3.4.3.tgz
- :x: **acorn-5.7.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/paulius-valiunas/dotenv-expand/commit/76f2af744d235e83423a6b3d04a4f551c958641e">76f2af744d235e83423a6b3d04a4f551c958641e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"acorn","packageVersion":"5.7.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"lab:13.1.0;espree:3.4.3;acorn:5.7.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0042","vulnerabilityDetails":"acorn is vulnerable to REGEX DoS. A regex of the form /[x-\\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.","vulnerabilityUrl":"https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | WS-2020-0042 (High) detected in acorn-5.7.3.tgz - ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.3.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p>
<p>Path to dependency file: dotenv-expand/package.json</p>
<p>Path to vulnerable library: dotenv-expand/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- lab-13.1.0.tgz (Root Library)
- espree-3.4.3.tgz
- :x: **acorn-5.7.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/paulius-valiunas/dotenv-expand/commit/76f2af744d235e83423a6b3d04a4f551c958641e">76f2af744d235e83423a6b3d04a4f551c958641e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"acorn","packageVersion":"5.7.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"lab:13.1.0;espree:3.4.3;acorn:5.7.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.1.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0042","vulnerabilityDetails":"acorn is vulnerable to REGEX DoS. A regex of the form /[x-\\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.","vulnerabilityUrl":"https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_test | ws high detected in acorn tgz ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file dotenv expand package json path to vulnerable library dotenv expand node modules acorn package json dependency hierarchy lab tgz root library espree tgz x acorn tgz vulnerable library found in head commit a href found in base branch master vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree lab espree acorn isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser vulnerabilityurl | 0 |
348,308 | 31,496,177,459 | IssuesEvent | 2023-08-31 02:39:35 | rancher/rancher | https://api.github.com/repos/rancher/rancher | reopened | [BUG] Rancher will spam helm-operation pods until it dies when a deployment fails | kind/bug [zube]: To Test team/mapps | **Rancher Server Setup**
- Rancher version: 2.7.3
- Installation option (Docker install/Helm Chart):
- If Helm Chart, Kubernetes Cluster and version (RKE1, RKE2, k3s, EKS, etc):
- Proxy/Cert Details:
**Information about the Cluster**
- Kubernetes version: 1.23.15
- Cluster Type (Local/Downstream):
local
- If downstream, what type of cluster? (Custom/Imported or specify provider for Hosted/Infrastructure Provider):
<!--
* Custom = Running a docker command on a node
* Imported = Running kubectl apply onto an existing k8s cluster
* Hosted = EKS, GKE, AKS, etc
* Infrastructure Provider = Rancher provisioning the nodes using different node drivers (e.g. AWS, Digital Ocean, etc)
-->
**User Information**
- What is the role of the user logged in? (Admin/Cluster Owner/Cluster Member/Project Owner/Project Member/Custom)
- If custom, define the set of permissions:
**Describe the bug**
<!--A clear and concise description of what the bug is.-->
EKS-operator had a bug that would cause a panic for some users. This panic would then prompy rancher to spin up a new helm-operation pod and attempt to reinstall eks-operator. This happened so rapidly that the local cluster ran out of available ips.
**To Reproduce**
<!--Steps to reproduce the behavior-->
1. install chart with pod that causes a panic. I would create a new chart with an image that reliably does this since the eks-operator panic is not easy to repro on command.
2. It is possible that not just any chart can cause this, in that case you may need to modify a hosted operator chart's image, for eks-operator, gke-operator, aks-operator.
**Result**
Helm-operation pods are spun up rapidly causing a large growth in resource usage, and kubernetes to run out of available pod ips.
**Expected Result**
<!--A clear and concise description of what you expected to happen.-->
Helm-operation pods are spun up, if necessary, at a rate that is manageable for rancher.
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem.-->
**Additional context**
<!--Add any other context about the problem here.-->
| 1.0 | [BUG] Rancher will spam helm-operation pods until it dies when a deployment fails - **Rancher Server Setup**
- Rancher version: 2.7.3
- Installation option (Docker install/Helm Chart):
- If Helm Chart, Kubernetes Cluster and version (RKE1, RKE2, k3s, EKS, etc):
- Proxy/Cert Details:
**Information about the Cluster**
- Kubernetes version: 1.23.15
- Cluster Type (Local/Downstream):
local
- If downstream, what type of cluster? (Custom/Imported or specify provider for Hosted/Infrastructure Provider):
<!--
* Custom = Running a docker command on a node
* Imported = Running kubectl apply onto an existing k8s cluster
* Hosted = EKS, GKE, AKS, etc
* Infrastructure Provider = Rancher provisioning the nodes using different node drivers (e.g. AWS, Digital Ocean, etc)
-->
**User Information**
- What is the role of the user logged in? (Admin/Cluster Owner/Cluster Member/Project Owner/Project Member/Custom)
- If custom, define the set of permissions:
**Describe the bug**
<!--A clear and concise description of what the bug is.-->
EKS-operator had a bug that would cause a panic for some users. This panic would then prompy rancher to spin up a new helm-operation pod and attempt to reinstall eks-operator. This happened so rapidly that the local cluster ran out of available ips.
**To Reproduce**
<!--Steps to reproduce the behavior-->
1. install chart with pod that causes a panic. I would create a new chart with an image that reliably does this since the eks-operator panic is not easy to repro on command.
2. It is possible that not just any chart can cause this, in that case you may need to modify a hosted operator chart's image, for eks-operator, gke-operator, aks-operator.
**Result**
Helm-operation pods are spun up rapidly causing a large growth in resource usage, and kubernetes to run out of available pod ips.
**Expected Result**
<!--A clear and concise description of what you expected to happen.-->
Helm-operation pods are spun up, if necessary, at a rate that is manageable for rancher.
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem.-->
**Additional context**
<!--Add any other context about the problem here.-->
| test | rancher will spam helm operation pods until it dies when a deployment fails rancher server setup rancher version installation option docker install helm chart if helm chart kubernetes cluster and version eks etc proxy cert details information about the cluster kubernetes version cluster type local downstream local if downstream what type of cluster custom imported or specify provider for hosted infrastructure provider custom running a docker command on a node imported running kubectl apply onto an existing cluster hosted eks gke aks etc infrastructure provider rancher provisioning the nodes using different node drivers e g aws digital ocean etc user information what is the role of the user logged in admin cluster owner cluster member project owner project member custom if custom define the set of permissions describe the bug eks operator had a bug that would cause a panic for some users this panic would then prompy rancher to spin up a new helm operation pod and attempt to reinstall eks operator this happened so rapidly that the local cluster ran out of available ips to reproduce install chart with pod that causes a panic i would create a new chart with an image that reliably does this since the eks operator panic is not easy to repro on command it is possible that not just any chart can cause this in that case you may need to modify a hosted operator chart s image for eks operator gke operator aks operator result helm operation pods are spun up rapidly causing a large growth in resource usage and kubernetes to run out of available pod ips expected result helm operation pods are spun up if necessary at a rate that is manageable for rancher screenshots additional context | 1 |
48,339 | 5,954,100,583 | IssuesEvent | 2017-05-27 14:40:23 | FireFly-WoW/FireFly-IssueTracker | https://api.github.com/repos/FireFly-WoW/FireFly-IssueTracker | closed | [Quest] Rescue OOX-22/FE! | Status: Tested / Working | **Description:**
Quest (quest=25476) won't complete when hitting the road, the chicken just keeps on running west.
**Current behaviour:**
Escort, test output, ambush spawn works perfetly but you wont get a complete message when you hit the road.
**Expected behaviour:**
The Quest should complete after the ambush once you hit the road. see video.
**Steps to reproduce the problem:**
1. Get the Quest.
2. Escort the chicken, defend it against the ambush, see no completion message at the road.
**Video:**
https://www.youtube.com/watch?v=DAuJLGV3TUk
| 1.0 | [Quest] Rescue OOX-22/FE! - **Description:**
Quest (quest=25476) won't complete when hitting the road, the chicken just keeps on running west.
**Current behaviour:**
Escort, test output, ambush spawn works perfetly but you wont get a complete message when you hit the road.
**Expected behaviour:**
The Quest should complete after the ambush once you hit the road. see video.
**Steps to reproduce the problem:**
1. Get the Quest.
2. Escort the chicken, defend it against the ambush, see no completion message at the road.
**Video:**
https://www.youtube.com/watch?v=DAuJLGV3TUk
| test | rescue oox fe description quest quest won t complete when hitting the road the chicken just keeps on running west current behaviour escort test output ambush spawn works perfetly but you wont get a complete message when you hit the road expected behaviour the quest should complete after the ambush once you hit the road see video steps to reproduce the problem get the quest escort the chicken defend it against the ambush see no completion message at the road video | 1 |
161,092 | 12,530,503,090 | IssuesEvent | 2020-06-04 13:11:23 | aces/Loris | https://api.github.com/repos/aces/Loris | closed | [Genomic Browser] Subproject Column displayed as Integer instead of Label | 23.0.0-testing Bug | **Describe the bug**
A. In the Profiles tab of the Genomic Browser, when downloading the Table as a CSV, the SubProject is displayed as an integer instead of a string label.
B. In the SNP tab of the Genomic Browser, the SubProject is displayed as an integer in the table instead of a string label
Test Plan Reference: https://github.com/aces/Loris/blob/master/modules/genomic_browser/test/TestPlan.md#data-download
**To Reproduce**
Steps to reproduce the behaviour (attach screenshots if applicable):
A.
1. Go to Profiles Tab of Genomic Browser
2. Click on Download as CSV.
3. Look at Subproject column of downloaded CSV.
B.
1. Go to SNP Tab of Genomic Browser
2. Look at Subproject column datatable.
A.

B.

| 1.0 | [Genomic Browser] Subproject Column displayed as Integer instead of Label - **Describe the bug**
A. In the Profiles tab of the Genomic Browser, when downloading the Table as a CSV, the SubProject is displayed as an integer instead of a string label.
B. In the SNP tab of the Genomic Browser, the SubProject is displayed as an integer in the table instead of a string label
Test Plan Reference: https://github.com/aces/Loris/blob/master/modules/genomic_browser/test/TestPlan.md#data-download
**To Reproduce**
Steps to reproduce the behaviour (attach screenshots if applicable):
A.
1. Go to Profiles Tab of Genomic Browser
2. Click on Download as CSV.
3. Look at Subproject column of downloaded CSV.
B.
1. Go to SNP Tab of Genomic Browser
2. Look at Subproject column datatable.
A.

B.

| test | subproject column displayed as integer instead of label describe the bug a in the profiles tab of the genomic browser when downloading the table as a csv the subproject is displayed as an integer instead of a string label b in the snp tab of the genomic browser the subproject is displayed as an integer in the table instead of a string label test plan reference to reproduce steps to reproduce the behaviour attach screenshots if applicable a go to profiles tab of genomic browser click on download as csv look at subproject column of downloaded csv b go to snp tab of genomic browser look at subproject column datatable a b | 1 |
130,829 | 10,661,465,323 | IssuesEvent | 2019-10-18 12:25:35 | CGAL/cgal | https://api.github.com/repos/CGAL/cgal | closed | inconsistent shortest path with CGAL/Surface_mesh_shortest_path | Pkg::Surface_mesh_shortest_path bug | I've done some testing with the [CGAL shortest path algorithm](https://doc.cgal.org/latest/Surface_mesh_shortest_path/index.html#Chapter_Surface_mesh_shortest_path) and I've come across an inconsistent result. The minimal code producing the wrong path can be downloaded [here](https://www.dropbox.com/s/k914a3vefvafr7o/cgal_discrete_geodesics_minimal_example2.tar.gz?dl=0). The archive contains a simple mesh in OFF format and a main.cpp (compilation via `g++ -frounding-math main.cpp -lCGAL -lgmp`). The minimal program requires the indices of the source and target point as arguments, the error is triggered with the arguments 4 and 11. The archive furthermore contains an image showing the mesh with the inconsistent path for the mentioned indices.
| 1.0 | inconsistent shortest path with CGAL/Surface_mesh_shortest_path - I've done some testing with the [CGAL shortest path algorithm](https://doc.cgal.org/latest/Surface_mesh_shortest_path/index.html#Chapter_Surface_mesh_shortest_path) and I've come across an inconsistent result. The minimal code producing the wrong path can be downloaded [here](https://www.dropbox.com/s/k914a3vefvafr7o/cgal_discrete_geodesics_minimal_example2.tar.gz?dl=0). The archive contains a simple mesh in OFF format and a main.cpp (compilation via `g++ -frounding-math main.cpp -lCGAL -lgmp`). The minimal program requires the indices of the source and target point as arguments, the error is triggered with the arguments 4 and 11. The archive furthermore contains an image showing the mesh with the inconsistent path for the mentioned indices.
| test | inconsistent shortest path with cgal surface mesh shortest path i ve done some testing with the and i ve come across an inconsistent result the minimal code producing the wrong path can be downloaded the archive contains a simple mesh in off format and a main cpp compilation via g frounding math main cpp lcgal lgmp the minimal program requires the indices of the source and target point as arguments the error is triggered with the arguments and the archive furthermore contains an image showing the mesh with the inconsistent path for the mentioned indices | 1 |
194,818 | 14,689,268,306 | IssuesEvent | 2021-01-02 08:38:48 | DCSFlightpanels/dcs-bios | https://api.github.com/repos/DCSFlightpanels/dcs-bios | closed | L-39ZA - FRONT_BOMB_REL_SEL 0 Puts Switch In Intermediate Position Instead of Down | Need testing-fix uploaded | FRONT_BOMB_REL_SEL 0 puts the switch in the position shown in the picture instead of full down. FRONT_BOMB_REL_SEL 1 puts the switch up (which is correct).

It may be an animation argument error. | 1.0 | L-39ZA - FRONT_BOMB_REL_SEL 0 Puts Switch In Intermediate Position Instead of Down - FRONT_BOMB_REL_SEL 0 puts the switch in the position shown in the picture instead of full down. FRONT_BOMB_REL_SEL 1 puts the switch up (which is correct).

It may be an animation argument error. | test | l front bomb rel sel puts switch in intermediate position instead of down front bomb rel sel puts the switch in the position shown in the picture instead of full down front bomb rel sel puts the switch up which is correct it may be an animation argument error | 1 |
454,009 | 13,093,441,394 | IssuesEvent | 2020-08-03 10:23:17 | SkriptLang/Skript | https://api.github.com/repos/SkriptLang/Skript | closed | ExprScripts: PatternSyntaxException | bug completed priority: medium | ### Description
An error is displayed in the console when I use the expression `scripts`.
### Steps to Reproduce
Just run `send "%scripts%"`.
### Expected Behavior
There should be no error.
### Errors / Screenshots
https://gist.github.com/Romitou/f7685d189b8735eec9e500b56d530fad
### Server Information
* **Server version/platform:** git-Spigot-800b93f-8160e29 (MC: 1.15.2)
* **Skript version:** tested with 2.5-alpha5 and 2.5-alpha6 (I forgot to test with 2.5-alpha6(rev1), but the error is finally the same).
| 1.0 | ExprScripts: PatternSyntaxException - ### Description
An error is displayed in the console when I use the expression `scripts`.
### Steps to Reproduce
Just run `send "%scripts%"`.
### Expected Behavior
There should be no error.
### Errors / Screenshots
https://gist.github.com/Romitou/f7685d189b8735eec9e500b56d530fad
### Server Information
* **Server version/platform:** git-Spigot-800b93f-8160e29 (MC: 1.15.2)
* **Skript version:** tested with 2.5-alpha5 and 2.5-alpha6 (I forgot to test with 2.5-alpha6(rev1), but the error is finally the same).
| non_test | exprscripts patternsyntaxexception description an error is displayed in the console when i use the expression scripts steps to reproduce just run send scripts expected behavior there should be no error errors screenshots server information server version platform git spigot mc skript version tested with and i forgot to test with but the error is finally the same | 0 |
98,475 | 8,678,130,289 | IssuesEvent | 2018-11-30 18:55:59 | backdrop/backdrop-issues | https://api.github.com/repos/backdrop/backdrop-issues | closed | [UX] The submit button in the "Add image style" page not styled as primary button. | pr - reviewed & tested by the community status - has pull request type - bug report | ### Steps to reproduce (if reporting a bug)
Navigate to `/admin/config/media/image-styles/add`...
### Actual behavior (if reporting a bug)
<img width="535" alt="screen shot 2018-11-05 at 7 15 20 pm" src="https://user-images.githubusercontent.com/2423362/47985952-5c1e4400-e12f-11e8-8e3a-e2831174fc21.png">
Hovered over:
<img width="536" alt="screen shot 2018-11-05 at 7 15 29 pm" src="https://user-images.githubusercontent.com/2423362/47985966-66d8d900-e12f-11e8-9bc3-07126d48999a.png">
### Expected behavior (if reporting a bug)
The button should be styled same as the rest of the primary submit buttons in the admin UI.
---
PR by @klonos: https://github.com/backdrop/backdrop/pull/2354 | 1.0 | [UX] The submit button in the "Add image style" page not styled as primary button. - ### Steps to reproduce (if reporting a bug)
Navigate to `/admin/config/media/image-styles/add`...
### Actual behavior (if reporting a bug)
<img width="535" alt="screen shot 2018-11-05 at 7 15 20 pm" src="https://user-images.githubusercontent.com/2423362/47985952-5c1e4400-e12f-11e8-8e3a-e2831174fc21.png">
Hovered over:
<img width="536" alt="screen shot 2018-11-05 at 7 15 29 pm" src="https://user-images.githubusercontent.com/2423362/47985966-66d8d900-e12f-11e8-9bc3-07126d48999a.png">
### Expected behavior (if reporting a bug)
The button should be styled same as the rest of the primary submit buttons in the admin UI.
---
PR by @klonos: https://github.com/backdrop/backdrop/pull/2354 | test | the submit button in the add image style page not styled as primary button steps to reproduce if reporting a bug navigate to admin config media image styles add actual behavior if reporting a bug img width alt screen shot at pm src hovered over img width alt screen shot at pm src expected behavior if reporting a bug the button should be styled same as the rest of the primary submit buttons in the admin ui pr by klonos | 1 |
163,898 | 20,364,249,518 | IssuesEvent | 2022-02-21 02:24:56 | Rossb0b/ecom-front | https://api.github.com/repos/Rossb0b/ecom-front | opened | CVE-2022-0512 (High) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2022-0512 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.901.12.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6.
<p>Publish Date: 2022-02-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p>
<p>Release Date: 2022-02-14</p>
<p>Fix Resolution: url-parse - 1.5.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-0512 (High) detected in url-parse-1.4.7.tgz - ## CVE-2022-0512 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.901.12.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6.
<p>Publish Date: 2022-02-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p>
<p>Release Date: 2022-02-14</p>
<p>Fix Resolution: url-parse - 1.5.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in url parse tgz cve high severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file package json path to vulnerable library node modules url parse package json dependency hierarchy build angular tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in base branch dev vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource | 0 |
368,074 | 25,776,035,797 | IssuesEvent | 2022-12-09 12:04:19 | bounswe/bounswe2022group4 | https://api.github.com/repos/bounswe/bounswe2022group4 | closed | Milestone-2 Report: Individual Report | Category - Documentation Priority - Medium Status: In Progress whom: individual | Description:
Individual contribution report need to be documented.
Steps:
1) Fill body of individual report sections
2) Link Issues and related PRs
Deadline: 09.12.2022 23.59
| 1.0 | Milestone-2 Report: Individual Report - Description:
Individual contribution report need to be documented.
Steps:
1) Fill body of individual report sections
2) Link Issues and related PRs
Deadline: 09.12.2022 23.59
| non_test | milestone report individual report description individual contribution report need to be documented steps fill body of individual report sections link issues and related prs deadline | 0 |
128,531 | 10,541,798,639 | IssuesEvent | 2019-10-02 11:43:41 | stylelint/stylelint | https://api.github.com/repos/stylelint/stylelint | closed | Improve test stability | status: wip type: tests | There are some tests that fail occasionally on Windows. It doesn't happen every time and not the same tests (it's always one of mentioned below).
First let's gather what failing tests we get overtime, and then find the problem and fix it.
Below will be mentioned Node.js version where test failed, and a link to result if possible. I don't think failures depend on Node.js version, though.
Node.js 12:
```
FAIL lib/__tests__/ignore.test.js
● using ignoreFiles with input files that would cause a postcss syntax error › syntax-error-ignored.scss found
expect(received).not.toBe(expected) // Object.is equality
Expected: not -1
313 | const index = Number(results[1].source.indexOf('syntax-error-ignored.scss'));
314 |
> 315 | expect(index).not.toBe(-1);
| ^
316 | });
317 |
318 | it('syntax-error-ignored.scss not linted', () => {
at Object.toBe (lib/__tests__/ignore.test.js:315:21)
```
[Node.js 8](https://github.com/stylelint/stylelint/runs/233974997):
```
FAIL lib/__tests__/ignore.test.js
● using ignoreFiles with input files that would cause a postcss syntax error › no-syntax-error.css found
expect(received).not.toBe(expected) // Object.is equality
Expected: not -1
302 |
303 | it('no-syntax-error.css found', () => {
> 304 | expect(results[0].source.indexOf('no-syntax-error.css')).not.toBe(-1);
| ^
305 | });
306 |
307 | it('no-syntax-error.css linted', () => {
at Object.toBe (lib/__tests__/ignore.test.js:304:64)
```
Node.js 12:
```
FAIL lib/__tests__/standalone-syntax.test.js
● standalone with postcss-safe-parser
expect(received).toBeFalsy()
Received: true
305 | const root = result._postcssResult.root;
306 |
> 307 | expect(results[0].errored).toBeFalsy();
| ^
308 | expect(results[0].warnings).toHaveLength(0);
309 | expect(root.toString()).not.toBe(root.source.input.css);
310 |
at toBeFalsy (lib/__tests__/standalone-syntax.test.js:307:33)
at Array.map (<anonymous>)
at map (lib/__tests__/standalone-syntax.test.js:303:12)
```
Node.js 8:
```
FAIL lib/__tests__/ignore.test.js
● using ignoreFiles with input files that would cause a postcss syntax error › syntax-error-ignored.scss marked as ignored
expect(received).toBeTruthy()
Received: undefined
321 |
322 | it('syntax-error-ignored.scss marked as ignored', () => {
> 323 | expect(results[1].ignored).toBeTruthy();
| ^
324 | });
325 | });
326 |
at Object.toBeTruthy (lib/__tests__/ignore.test.js:323:30)
``` | 1.0 | Improve test stability - There are some tests that fail occasionally on Windows. It doesn't happen every time and not the same tests (it's always one of mentioned below).
First let's gather what failing tests we get overtime, and then find the problem and fix it.
Below will be mentioned Node.js version where test failed, and a link to result if possible. I don't think failures depend on Node.js version, though.
Node.js 12:
```
FAIL lib/__tests__/ignore.test.js
● using ignoreFiles with input files that would cause a postcss syntax error › syntax-error-ignored.scss found
expect(received).not.toBe(expected) // Object.is equality
Expected: not -1
313 | const index = Number(results[1].source.indexOf('syntax-error-ignored.scss'));
314 |
> 315 | expect(index).not.toBe(-1);
| ^
316 | });
317 |
318 | it('syntax-error-ignored.scss not linted', () => {
at Object.toBe (lib/__tests__/ignore.test.js:315:21)
```
[Node.js 8](https://github.com/stylelint/stylelint/runs/233974997):
```
FAIL lib/__tests__/ignore.test.js
● using ignoreFiles with input files that would cause a postcss syntax error › no-syntax-error.css found
expect(received).not.toBe(expected) // Object.is equality
Expected: not -1
302 |
303 | it('no-syntax-error.css found', () => {
> 304 | expect(results[0].source.indexOf('no-syntax-error.css')).not.toBe(-1);
| ^
305 | });
306 |
307 | it('no-syntax-error.css linted', () => {
at Object.toBe (lib/__tests__/ignore.test.js:304:64)
```
Node.js 12:
```
FAIL lib/__tests__/standalone-syntax.test.js
● standalone with postcss-safe-parser
expect(received).toBeFalsy()
Received: true
305 | const root = result._postcssResult.root;
306 |
> 307 | expect(results[0].errored).toBeFalsy();
| ^
308 | expect(results[0].warnings).toHaveLength(0);
309 | expect(root.toString()).not.toBe(root.source.input.css);
310 |
at toBeFalsy (lib/__tests__/standalone-syntax.test.js:307:33)
at Array.map (<anonymous>)
at map (lib/__tests__/standalone-syntax.test.js:303:12)
```
Node.js 8:
```
FAIL lib/__tests__/ignore.test.js
● using ignoreFiles with input files that would cause a postcss syntax error › syntax-error-ignored.scss marked as ignored
expect(received).toBeTruthy()
Received: undefined
321 |
322 | it('syntax-error-ignored.scss marked as ignored', () => {
> 323 | expect(results[1].ignored).toBeTruthy();
| ^
324 | });
325 | });
326 |
at Object.toBeTruthy (lib/__tests__/ignore.test.js:323:30)
``` | test | improve test stability there are some tests that fail occasionally on windows it doesn t happen every time and not the same tests it s always one of mentioned below first let s gather what failing tests we get overtime and then find the problem and fix it below will be mentioned node js version where test failed and a link to result if possible i don t think failures depend on node js version though node js fail lib tests ignore test js ● using ignorefiles with input files that would cause a postcss syntax error › syntax error ignored scss found expect received not tobe expected object is equality expected not const index number results source indexof syntax error ignored scss expect index not tobe it syntax error ignored scss not linted at object tobe lib tests ignore test js fail lib tests ignore test js ● using ignorefiles with input files that would cause a postcss syntax error › no syntax error css found expect received not tobe expected object is equality expected not it no syntax error css found expect results source indexof no syntax error css not tobe it no syntax error css linted at object tobe lib tests ignore test js node js fail lib tests standalone syntax test js ● standalone with postcss safe parser expect received tobefalsy received true const root result postcssresult root expect results errored tobefalsy expect results warnings tohavelength expect root tostring not tobe root source input css at tobefalsy lib tests standalone syntax test js at array map at map lib tests standalone syntax test js node js fail lib tests ignore test js ● using ignorefiles with input files that would cause a postcss syntax error › syntax error ignored scss marked as ignored expect received tobetruthy received undefined it syntax error ignored scss marked as ignored expect results ignored tobetruthy at object tobetruthy lib tests ignore test js | 1 |
65,722 | 16,459,396,227 | IssuesEvent | 2021-05-21 16:35:10 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | opened | GCB builder may exhaust `get` quota with large number of build artifacts | area/build build/gcb kind/feature-request priority/p2 source/partnerships | GCB has a quota of 660req/min on `get`s to check build results. When dealing with builds with many artifacts, with a larger team, it's not impossible to hit this threshold.
We currently have [an exponential backoff](https://github.com/GoogleContainerTools/skaffold/blob/ff5038a51e8408e5ac80f66b6867ff2b4ba528c2/pkg/skaffold/build/gcb/types.go#L71-L78) that starts at 1s and increases by 1.5x per step to a maximum of 10 steps with a cap of 60s. That means we could check 8 times per minute per container. With 15 containers in the project, that could be hit if 5 people are building simultaneously (e.g., incl CI builds).
GCB supports a [`list` operation](https://cloud.google.com/build/docs/api/reference/rest/v1/projects.builds/list) that takes a filter. We should be able to use this operation for the suite of builds that have been created and reduce our API use.
| 2.0 | GCB builder may exhaust `get` quota with large number of build artifacts - GCB has a quota of 660req/min on `get`s to check build results. When dealing with builds with many artifacts, with a larger team, it's not impossible to hit this threshold.
We currently have [an exponential backoff](https://github.com/GoogleContainerTools/skaffold/blob/ff5038a51e8408e5ac80f66b6867ff2b4ba528c2/pkg/skaffold/build/gcb/types.go#L71-L78) that starts at 1s and increases by 1.5x per step to a maximum of 10 steps with a cap of 60s. That means we could check 8 times per minute per container. With 15 containers in the project, that could be hit if 5 people are building simultaneously (e.g., incl CI builds).
GCB supports a [`list` operation](https://cloud.google.com/build/docs/api/reference/rest/v1/projects.builds/list) that takes a filter. We should be able to use this operation for the suite of builds that have been created and reduce our API use.
| non_test | gcb builder may exhaust get quota with large number of build artifacts gcb has a quota of min on get s to check build results when dealing with builds with many artifacts with a larger team it s not impossible to hit this threshold we currently have that starts at and increases by per step to a maximum of steps with a cap of that means we could check times per minute per container with containers in the project that could be hit if people are building simultaneously e g incl ci builds gcb supports a that takes a filter we should be able to use this operation for the suite of builds that have been created and reduce our api use | 0 |
830,137 | 31,991,373,574 | IssuesEvent | 2023-09-21 06:10:00 | lmareksla/DPE_Issues | https://api.github.com/repos/lmareksla/DPE_Issues | opened | Hist - weird steps | bug question middle priority | ### feature/issue description
Found weird steps in the distribution:

### program and data specification
**DPE version:** 1.1.0 230919 33b18fc9
**data type:** t3pa
**used settings of DPE:** standard
**pc configuration:** ubuntu22
### issue originator
LM
### how to reproduce
pregenerated data with constant count of particles in each sample
### solution description
?
### tests
test_050
### issue process
- [x] reproduced
- [ ] solution / root cause discovered
- [ ] solution /fix implemented
- [ ] tested
| 1.0 | Hist - weird steps - ### feature/issue description
Found weird steps in the distribution:

### program and data specification
**DPE version:** 1.1.0 230919 33b18fc9
**data type:** t3pa
**used settings of DPE:** standard
**pc configuration:** ubuntu22
### issue originator
LM
### how to reproduce
pregenerated data with constant count of particles in each sample
### solution description
?
### tests
test_050
### issue process
- [x] reproduced
- [ ] solution / root cause discovered
- [ ] solution /fix implemented
- [ ] tested
| non_test | hist weird steps feature issue description found weird steps in the distribution program and data specification dpe version data type used settings of dpe standard pc configuration issue originator lm how to reproduce pregenerated data with constant count of particles in each sample solution description tests test issue process reproduced solution root cause discovered solution fix implemented tested | 0 |
727,332 | 25,031,823,518 | IssuesEvent | 2022-11-04 13:02:43 | microsoft/PowerToys | https://api.github.com/repos/microsoft/PowerToys | closed | Installer kills powertoys.exe without verifying it is installer itself | Issue-Bug Needs-Author-Feedback Area-Setup/Install Priority-2 Status-No recent activity | ### Microsoft PowerToys version
0.37.2
### Running as admin
- [X] Yes
### Area(s) with issue?
Installer
### Steps to reproduce
Install
### ✔️ Expected Behavior
Install
### ❌ Actual Behavior
Quits right after I click install. I'm currently running the 3.1.13 runtime that it comes with, but I've also tried it with the recommended .NET core recommended [here](https://docs.microsoft.com/en-us/windows/powertoys/install#requirements).
### Other Software
_No response_ | 1.0 | Installer kills powertoys.exe without verifying it is installer itself - ### Microsoft PowerToys version
0.37.2
### Running as admin
- [X] Yes
### Area(s) with issue?
Installer
### Steps to reproduce
Install
### ✔️ Expected Behavior
Install
### ❌ Actual Behavior
Quits right after I click install. I'm currently running the 3.1.13 runtime that it comes with, but I've also tried it with the recommended .NET core recommended [here](https://docs.microsoft.com/en-us/windows/powertoys/install#requirements).
### Other Software
_No response_ | non_test | installer kills powertoys exe without verifying it is installer itself microsoft powertoys version running as admin yes area s with issue installer steps to reproduce install ✔️ expected behavior install ❌ actual behavior quits right after i click install i m currently running the runtime that it comes with but i ve also tried it with the recommended net core recommended other software no response | 0 |
46,674 | 5,826,742,455 | IssuesEvent | 2017-05-08 06:29:47 | Kademi/kademi-dev | https://api.github.com/repos/Kademi/kademi-dev | closed | RewardStoreApp menu items are not using the points admin role | bug Ready to Test - Dev Ready to Test QA | Currently the RewardStoreApp is using the content author permission to decide whether or not to show menu items:
```
if ( canAccessAdminContent(parent) ) {
parent.getOrCreate("menuManageRewardStore", " Reward stores", parentPath.child("reward-store")).setOrdering(
110);
parent.getOrCreate("menuManagePointsSystems", "Points buckets", parentPath.child("points-buckets")).setOrdering(
120);
parent.getOrCreate("menuManagePointsAllocationRules", "Points Allocation Rules", parentPath.child("points-rules")).setOrdering(
130);
parent.getOrCreate("menuManagePointsStatements", "Points Statements", parentPath.child("pointsBalanceStats")).setOrdering(
140);
}
....
private boolean canAccessAdminContent(MenuItem parent) {
return parent.getOrgRoles().contains(adminRole)
|| parent.getOrgRoles().contains(ContentApp.CONTENT_AUTHOR_ROLE);
}
```
But this is incorrect. The app should check for the presence of the PointsAdministratorRole, not content author role.
| 2.0 | RewardStoreApp menu items are not using the points admin role - Currently the RewardStoreApp is using the content author permission to decide whether or not to show menu items:
```
if ( canAccessAdminContent(parent) ) {
parent.getOrCreate("menuManageRewardStore", " Reward stores", parentPath.child("reward-store")).setOrdering(
110);
parent.getOrCreate("menuManagePointsSystems", "Points buckets", parentPath.child("points-buckets")).setOrdering(
120);
parent.getOrCreate("menuManagePointsAllocationRules", "Points Allocation Rules", parentPath.child("points-rules")).setOrdering(
130);
parent.getOrCreate("menuManagePointsStatements", "Points Statements", parentPath.child("pointsBalanceStats")).setOrdering(
140);
}
....
private boolean canAccessAdminContent(MenuItem parent) {
return parent.getOrgRoles().contains(adminRole)
|| parent.getOrgRoles().contains(ContentApp.CONTENT_AUTHOR_ROLE);
}
```
But this is incorrect. The app should check for the presence of the PointsAdministratorRole, not content author role.
| test | rewardstoreapp menu items are not using the points admin role currently the rewardstoreapp is using the content author permission to decide whether or not to show menu items if canaccessadmincontent parent parent getorcreate menumanagerewardstore reward stores parentpath child reward store setordering parent getorcreate menumanagepointssystems points buckets parentpath child points buckets setordering parent getorcreate menumanagepointsallocationrules points allocation rules parentpath child points rules setordering parent getorcreate menumanagepointsstatements points statements parentpath child pointsbalancestats setordering private boolean canaccessadmincontent menuitem parent return parent getorgroles contains adminrole parent getorgroles contains contentapp content author role but this is incorrect the app should check for the presence of the pointsadministratorrole not content author role | 1 |
174,913 | 6,544,366,150 | IssuesEvent | 2017-09-03 15:21:02 | mentdotio/ment.io | https://api.github.com/repos/mentdotio/ment.io | closed | Use Postgraphile as server plugin | low priority | The CLI option for postgraphile works wonderfully and _is_ currently sufficient, however the running the server app **should** work, and it currently doesn't. | 1.0 | Use Postgraphile as server plugin - The CLI option for postgraphile works wonderfully and _is_ currently sufficient, however the running the server app **should** work, and it currently doesn't. | non_test | use postgraphile as server plugin the cli option for postgraphile works wonderfully and is currently sufficient however the running the server app should work and it currently doesn t | 0 |
2,311 | 7,633,728,970 | IssuesEvent | 2018-05-06 09:26:08 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Failed to register node | kind/bug lifecycle/rotten sig/architecture sig/node | **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
**What happened**:
If instance comes from the zone which has Chinese name,it can not be used as a kubernetes node.
\-----------
\# nova show ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka | grep availability_zone
| OS-EXT-AZ:availability_zone | 中文 |
\-----------
Kubelet's log as following:
\-----------
Jul 20 07:35:56 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: I0720 07:35:56.762145 10092 kubelet_node_status.go:283] Adding node label from cloud provider: failure-domain.beta.kubernetes.io/zone=中文
Jul 20 07:35:56 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: I0720 07:35:56.762202 10092 kubelet_node_status.go:287] Adding node label from cloud provider: failure-domain.beta.kubernetes.io/region=RegionOne
Jul 20 07:35:57 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: I0720 07:35:57.527723 10092 kubelet_node_status.go:77] Attempting to register node ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka
Jul 20 07:35:57 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: E0720 07:35:57.530644 10092 kubelet_node_status.go:101] Unable to register node "ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka" with API server: Node "ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka" is invalid: metadata.labels: Invalid value: "中文": a valid label must be an empty string or consist of alphanumeric characters, '-', '_' or '.', and must start and end with an alphanumeric character (e.g. 'MyValue', or 'my_value', or '12345', regex used for validation is '(([A-Za-z0-9][-A-Za-z0-9_.]*)?[A-Za-z0-9])?')
Jul 20 07:35:58 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal
\-----------
**What you expected to happen**:
Kubernetes node also can comes from Chinese name zone.
**How to reproduce it (as minimally and precisely as possible)**:
Create a instance in Chinese-name-zone, then start kubelet and register to kube-apiserver.
**Anything else we need to know?**:
**Environment**:
- Kubernetes version (use `kubectl version`): v1.6.4
- Cloud provider or hardware configuration**: OpenStack cloud provider
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
| 1.0 | Failed to register node - **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
**What happened**:
If instance comes from the zone which has Chinese name,it can not be used as a kubernetes node.
\-----------
\# nova show ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka | grep availability_zone
| OS-EXT-AZ:availability_zone | 中文 |
\-----------
Kubelet's log as following:
\-----------
Jul 20 07:35:56 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: I0720 07:35:56.762145 10092 kubelet_node_status.go:283] Adding node label from cloud provider: failure-domain.beta.kubernetes.io/zone=中文
Jul 20 07:35:56 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: I0720 07:35:56.762202 10092 kubelet_node_status.go:287] Adding node label from cloud provider: failure-domain.beta.kubernetes.io/region=RegionOne
Jul 20 07:35:57 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: I0720 07:35:57.527723 10092 kubelet_node_status.go:77] Attempting to register node ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka
Jul 20 07:35:57 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal kubelet[10092]: E0720 07:35:57.530644 10092 kubelet_node_status.go:101] Unable to register node "ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka" with API server: Node "ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka" is invalid: metadata.labels: Invalid value: "中文": a valid label must be an empty string or consist of alphanumeric characters, '-', '_' or '.', and must start and end with an alphanumeric character (e.g. 'MyValue', or 'my_value', or '12345', regex used for validation is '(([A-Za-z0-9][-A-Za-z0-9_.]*)?[A-Za-z0-9])?')
Jul 20 07:35:58 ne-cdlqmo7wgw-0-7yzmth4qkivz-kube-minion-btgwnwp3utka.novalocal
\-----------
**What you expected to happen**:
Kubernetes node also can comes from Chinese name zone.
**How to reproduce it (as minimally and precisely as possible)**:
Create a instance in Chinese-name-zone, then start kubelet and register to kube-apiserver.
**Anything else we need to know?**:
**Environment**:
- Kubernetes version (use `kubectl version`): v1.6.4
- Cloud provider or hardware configuration**: OpenStack cloud provider
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
| non_test | failed to register node is this a bug report or feature request kind bug what happened if instance comes from the zone which has chinese name,it can not be used as a kubernetes node nova show ne kube minion grep availability zone os ext az availability zone 中文 kubelet s log as following jul ne kube minion novalocal kubelet kubelet node status go adding node label from cloud provider failure domain beta kubernetes io zone 中文 jul ne kube minion novalocal kubelet kubelet node status go adding node label from cloud provider failure domain beta kubernetes io region regionone jul ne kube minion novalocal kubelet kubelet node status go attempting to register node ne kube minion jul ne kube minion novalocal kubelet kubelet node status go unable to register node ne kube minion with api server node ne kube minion is invalid metadata labels invalid value 中文 a valid label must be an empty string or consist of alphanumeric characters or and must start and end with an alphanumeric character e g myvalue or my value or regex used for validation is jul ne kube minion novalocal what you expected to happen kubernetes node also can comes from chinese name zone how to reproduce it as minimally and precisely as possible create a instance in chinese name zone then start kubelet and register to kube apiserver anything else we need to know environment kubernetes version use kubectl version cloud provider or hardware configuration openstack cloud provider os e g from etc os release kernel e g uname a install tools others | 0 |
3,265 | 3,102,823,853 | IssuesEvent | 2015-08-31 04:03:42 | turnkeylinux/tracker | https://api.github.com/repos/turnkeylinux/tracker | closed | [Jessie] OpenLDAP appliance not building | bug build inprogress openldap | ````
[ ok ] Starting OpenLDAP: slapd.
Processing triggers for libc-bin (2.19-18) ...
adding new entry "ou=Groups,dc=example,dc=com"
adding new entry "cn=default,ou=Groups,dc=example,dc=com"
adding new entry "ou=Users,dc=example,dc=com"
SASL/EXTERNAL authentication started
SASL username: gidNumber=0+uidNumber=0,cn=peercred,cn=external,cn=auth
SASL SSF: 0
modifying entry "cn=config"
SASL/EXTERNAL authentication started
SASL username: gidNumber=0+uidNumber=0,cn=peercred,cn=external,cn=auth
SASL SSF: 0
modifying entry "olcDatabase={1}hdb,cn=config"
ldap_modify: Other (e.g., implementation specific) error (80)
additional info: duplicate index definition for attr "cn"
Traceback (most recent call last):
File "/usr/lib/inithooks/bin/openldap.py", line 66, in <module>
main()
File "/usr/lib/inithooks/bin/openldap.py", line 63, in main
system(script, domain, password)
File "/usr/lib/python2.7/dist-packages/executil.py", line 56, in system
raise ExecError(command, exitcode)
executil.ExecError: non-zero exitcode (80) for command: /usr/lib/inithooks/bin/openldap-reinit.sh 'example.com' 'turnkey'
/usr/share/fab/product.mk:476: recipe for target 'build/stamps/root.patched' failed
make: *** [build/stamps/root.patched] Error 1
````
Seems to be a config issue. Not sure whether it's because of Jessie or perhaps some more general issue. FWIW I tested building from the v13.0 tag and that builds ok. | 1.0 | [Jessie] OpenLDAP appliance not building - ````
[ ok ] Starting OpenLDAP: slapd.
Processing triggers for libc-bin (2.19-18) ...
adding new entry "ou=Groups,dc=example,dc=com"
adding new entry "cn=default,ou=Groups,dc=example,dc=com"
adding new entry "ou=Users,dc=example,dc=com"
SASL/EXTERNAL authentication started
SASL username: gidNumber=0+uidNumber=0,cn=peercred,cn=external,cn=auth
SASL SSF: 0
modifying entry "cn=config"
SASL/EXTERNAL authentication started
SASL username: gidNumber=0+uidNumber=0,cn=peercred,cn=external,cn=auth
SASL SSF: 0
modifying entry "olcDatabase={1}hdb,cn=config"
ldap_modify: Other (e.g., implementation specific) error (80)
additional info: duplicate index definition for attr "cn"
Traceback (most recent call last):
File "/usr/lib/inithooks/bin/openldap.py", line 66, in <module>
main()
File "/usr/lib/inithooks/bin/openldap.py", line 63, in main
system(script, domain, password)
File "/usr/lib/python2.7/dist-packages/executil.py", line 56, in system
raise ExecError(command, exitcode)
executil.ExecError: non-zero exitcode (80) for command: /usr/lib/inithooks/bin/openldap-reinit.sh 'example.com' 'turnkey'
/usr/share/fab/product.mk:476: recipe for target 'build/stamps/root.patched' failed
make: *** [build/stamps/root.patched] Error 1
````
Seems to be a config issue. Not sure whether it's because of Jessie or perhaps some more general issue. FWIW I tested building from the v13.0 tag and that builds ok. | non_test | openldap appliance not building starting openldap slapd processing triggers for libc bin adding new entry ou groups dc example dc com adding new entry cn default ou groups dc example dc com adding new entry ou users dc example dc com sasl external authentication started sasl username gidnumber uidnumber cn peercred cn external cn auth sasl ssf modifying entry cn config sasl external authentication started sasl username gidnumber uidnumber cn peercred cn external cn auth sasl ssf modifying entry olcdatabase hdb cn config ldap modify other e g implementation specific error additional info duplicate index definition for attr cn traceback most recent call last file usr lib inithooks bin openldap py line in main file usr lib inithooks bin openldap py line in main system script domain password file usr lib dist packages executil py line in system raise execerror command exitcode executil execerror non zero exitcode for command usr lib inithooks bin openldap reinit sh example com turnkey usr share fab product mk recipe for target build stamps root patched failed make error seems to be a config issue not sure whether it s because of jessie or perhaps some more general issue fwiw i tested building from the tag and that builds ok | 0 |
747,242 | 26,078,199,203 | IssuesEvent | 2022-12-24 22:30:54 | Thorfusion/Mekanism-1.7.10-Community-Edition | https://api.github.com/repos/Thorfusion/Mekanism-1.7.10-Community-Edition | closed | [BUG]The machine in MOD will not automatically close the GUI when the player is far away from the machine, so that you can brush items | bug PRIORITY: MED STATUS: RESEARCHING | **Describe the bug**
The machine in MOD will not automatically close the GUI when the player is far away from the machine, so that you can brush items
**To Reproduce**
Steps to reproduce the behavior:
1. Place any Machine Block and Place items that can move players
2. Put items in GUI
3. Standing on the conveyor belt
4. open gui
5. Wait to move out of the load chunk set by the server
6. Take out items,close gui
7. Run to the machine location and open it
8. See error
**Expected behavior**
Close the GUI when the player is a certain distance from the machine
**Screenshots**





**Desktop (please complete the following information):**
- OS: server
- Modpack: any
- Modpack Version: any
- Mekanism Version: 9.10.16
Note that we only provide cross-mod support in the official modpacks for this mod. We always offer help and we may fix issues in public packs not on curseforge.
**PROVIDE the following files**
+ The crash report in folder ./crash-reports (both server and client logs)
+ The full log located in the folder ./logs
+ The configfiles in ./config/mekanism/ if edited
+ If the issue is with multiblock structures or similar complex stuff, also provide the save file where the error occurs. So that our time is saved for other bugs instead of trying to copy your build.
Use gist.github.com to provide such files.
https://user-images.githubusercontent.com/99242646/205433355-6604b595-e91f-4d5b-8be9-8e7035ae6617.mp4
Issues not providing the required files will not get help
**Additional context**
Add any other context about the problem here.
| 1.0 | [BUG]The machine in MOD will not automatically close the GUI when the player is far away from the machine, so that you can brush items - **Describe the bug**
The machine in MOD will not automatically close the GUI when the player is far away from the machine, so that you can brush items
**To Reproduce**
Steps to reproduce the behavior:
1. Place any Machine Block and Place items that can move players
2. Put items in GUI
3. Standing on the conveyor belt
4. open gui
5. Wait to move out of the load chunk set by the server
6. Take out items,close gui
7. Run to the machine location and open it
8. See error
**Expected behavior**
Close the GUI when the player is a certain distance from the machine
**Screenshots**





**Desktop (please complete the following information):**
- OS: server
- Modpack: any
- Modpack Version: any
- Mekanism Version: 9.10.16
Note that we only provide cross-mod support in the official modpacks for this mod. We always offer help and we may fix issues in public packs not on curseforge.
**PROVIDE the following files**
+ The crash report in folder ./crash-reports (both server and client logs)
+ The full log located in the folder ./logs
+ The configfiles in ./config/mekanism/ if edited
+ If the issue is with multiblock structures or similar complex stuff, also provide the save file where the error occurs. So that our time is saved for other bugs instead of trying to copy your build.
Use gist.github.com to provide such files.
https://user-images.githubusercontent.com/99242646/205433355-6604b595-e91f-4d5b-8be9-8e7035ae6617.mp4
Issues not providing the required files will not get help
**Additional context**
Add any other context about the problem here.
| non_test | the machine in mod will not automatically close the gui when the player is far away from the machine so that you can brush items describe the bug the machine in mod will not automatically close the gui when the player is far away from the machine so that you can brush items to reproduce steps to reproduce the behavior place any machine block and place items that can move players put items in gui standing on the conveyor belt open gui wait to move out of the load chunk set by the server take out items close gui run to the machine location and open it see error expected behavior close the gui when the player is a certain distance from the machine screenshots desktop please complete the following information os server modpack any modpack version any mekanism version note that we only provide cross mod support in the official modpacks for this mod we always offer help and we may fix issues in public packs not on curseforge provide the following files the crash report in folder crash reports both server and client logs the full log located in the folder logs the configfiles in config mekanism if edited if the issue is with multiblock structures or similar complex stuff also provide the save file where the error occurs so that our time is saved for other bugs instead of trying to copy your build use gist github com to provide such files issues not providing the required files will not get help additional context add any other context about the problem here | 0 |
10,278 | 3,097,653,485 | IssuesEvent | 2015-08-28 04:27:23 | FreeCodeCamp/FreeCodeCamp | https://api.github.com/repos/FreeCodeCamp/FreeCodeCamp | closed | Failed test, Cannot read property 'length' of null when spacing encountered in solution | confirmed Discussing Test Improvement | Challenge http://www.freecodecamp.com/challenges/waypoint-override-all-other-styles-by-using-important has an issue.
If there is a space between "!important" and ";" test will return:
```
Cannot read property 'length' of null
```
Having a space between last character and ";" on other lines do not provide a failed test case.
```
<style>
body {
background-color: black;
font-family: Monospace;
color: green;
}
#orange-text {
color: orange;
}
.pink-text {
color: pink !important ;
}
.blue-text {
color: blue;
}
</style>
<h1 id="orange-text" class="pink-text blue-text" style="color: white">Hello World!</h1>
``` | 1.0 | Failed test, Cannot read property 'length' of null when spacing encountered in solution - Challenge http://www.freecodecamp.com/challenges/waypoint-override-all-other-styles-by-using-important has an issue.
If there is a space between "!important" and ";" test will return:
```
Cannot read property 'length' of null
```
Having a space between last character and ";" on other lines do not provide a failed test case.
```
<style>
body {
background-color: black;
font-family: Monospace;
color: green;
}
#orange-text {
color: orange;
}
.pink-text {
color: pink !important ;
}
.blue-text {
color: blue;
}
</style>
<h1 id="orange-text" class="pink-text blue-text" style="color: white">Hello World!</h1>
``` | test | failed test cannot read property length of null when spacing encountered in solution challenge has an issue if there is a space between important and test will return cannot read property length of null having a space between last character and on other lines do not provide a failed test case body background color black font family monospace color green orange text color orange pink text color pink important blue text color blue hello world | 1 |
96,420 | 8,614,284,048 | IssuesEvent | 2018-11-19 17:03:43 | SME-Issues/issues | https://api.github.com/repos/SME-Issues/issues | closed | Test Summary - 19/11/2018 - 5004 | NLP Api pulse_tests | ### Comprehension
- Compound Query Tests Invoice Partial (5): **100%** pass (5), 0 failed understood (-)
| 1.0 | Test Summary - 19/11/2018 - 5004 - ### Comprehension
- Compound Query Tests Invoice Partial (5): **100%** pass (5), 0 failed understood (-)
| test | test summary comprehension compound query tests invoice partial pass failed understood | 1 |
34,147 | 12,244,925,633 | IssuesEvent | 2020-05-05 12:05:11 | Souparnee/WebGoat2_20022020 | https://api.github.com/repos/Souparnee/WebGoat2_20022020 | opened | CVE-2019-6284 (Medium) detected in node-sass-4.11.0.tgz, opennms-opennms-source-22.0.1-1 | security vulnerability | ## CVE-2019-6284 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/WebGoat2_20022020/docs/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/WebGoat2_20022020/docs/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-4.0.2.tgz (Root Library)
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Souparnee/WebGoat2_20022020/commit/92d0ea06552607a1d3a17f8a21d12f430d4a4115">92d0ea06552607a1d3a17f8a21d12f430d4a4115</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-6284 (Medium) detected in node-sass-4.11.0.tgz, opennms-opennms-source-22.0.1-1 - ## CVE-2019-6284 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/WebGoat2_20022020/docs/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/WebGoat2_20022020/docs/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-4.0.2.tgz (Root Library)
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Souparnee/WebGoat2_20022020/commit/92d0ea06552607a1d3a17f8a21d12f430d4a4115">92d0ea06552607a1d3a17f8a21d12f430d4a4115</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6284</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in node sass tgz opennms opennms source cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm docs package json path to vulnerable library tmp ws scm docs node modules node sass package json dependency hierarchy gulp sass tgz root library x node sass tgz vulnerable library found in head commit a href vulnerability details in libsass a heap based buffer over read exists in sass prelexer alternatives in prelexer hpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
290,983 | 25,112,469,378 | IssuesEvent | 2022-11-08 21:50:58 | opensearch-project/OpenSearch | https://api.github.com/repos/opensearch-project/OpenSearch | opened | [BUG] testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections is flaky | bug flaky-test | **Describe the bug**
```
REPRODUCE WITH: ./gradlew ':server:test' --tests "org.opensearch.index.ShardIndexingPressureConcurrentExecutionTests.testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections" -Dtests.seed=3AA0904AC11EAEA5 -Dtests.security.manager=true -Dtests.jvm.argline="-XX:TieredStopAtLevel=1 -XX:ReservedCodeCacheSize=64m" -Dtests.locale=uk-UA -Dtests.timezone=America/Argentina/Jujuy -Druntime.java=19
org.opensearch.index.ShardIndexingPressureConcurrentExecutionTests > testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections FAILED
java.lang.AssertionError: expected null, but was:<org.opensearch.index.stats.IndexingPressurePerShardStats@3a0179b7>
at __randomizedtesting.SeedInfo.seed([3AA0904AC11EAEA5:BC95A9DD94E04AA8]:0)
at org.junit.Assert.fail(Assert.java:89)
at org.junit.Assert.failNotNull(Assert.java:756)
at org.junit.Assert.assertNull(Assert.java:738)
at org.junit.Assert.assertNull(Assert.java:748)
at org.opensearch.index.ShardIndexingPressureConcurrentExecutionTests.testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections(ShardIndexingPressureConcurrentExecutionTests.java:274)
```
https://github.com/opensearch-project/OpenSearch/pull/5143 | 1.0 | [BUG] testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections is flaky - **Describe the bug**
```
REPRODUCE WITH: ./gradlew ':server:test' --tests "org.opensearch.index.ShardIndexingPressureConcurrentExecutionTests.testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections" -Dtests.seed=3AA0904AC11EAEA5 -Dtests.security.manager=true -Dtests.jvm.argline="-XX:TieredStopAtLevel=1 -XX:ReservedCodeCacheSize=64m" -Dtests.locale=uk-UA -Dtests.timezone=America/Argentina/Jujuy -Druntime.java=19
org.opensearch.index.ShardIndexingPressureConcurrentExecutionTests > testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections FAILED
java.lang.AssertionError: expected null, but was:<org.opensearch.index.stats.IndexingPressurePerShardStats@3a0179b7>
at __randomizedtesting.SeedInfo.seed([3AA0904AC11EAEA5:BC95A9DD94E04AA8]:0)
at org.junit.Assert.fail(Assert.java:89)
at org.junit.Assert.failNotNull(Assert.java:756)
at org.junit.Assert.assertNull(Assert.java:738)
at org.junit.Assert.assertNull(Assert.java:748)
at org.opensearch.index.ShardIndexingPressureConcurrentExecutionTests.testCoordinatingPrimaryThreadedUpdateToShardLimitsAndRejections(ShardIndexingPressureConcurrentExecutionTests.java:274)
```
https://github.com/opensearch-project/OpenSearch/pull/5143 | test | testcoordinatingprimarythreadedupdatetoshardlimitsandrejections is flaky describe the bug reproduce with gradlew server test tests org opensearch index shardindexingpressureconcurrentexecutiontests testcoordinatingprimarythreadedupdatetoshardlimitsandrejections dtests seed dtests security manager true dtests jvm argline xx tieredstopatlevel xx reservedcodecachesize dtests locale uk ua dtests timezone america argentina jujuy druntime java org opensearch index shardindexingpressureconcurrentexecutiontests testcoordinatingprimarythreadedupdatetoshardlimitsandrejections failed java lang assertionerror expected null but was at randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert failnotnull assert java at org junit assert assertnull assert java at org junit assert assertnull assert java at org opensearch index shardindexingpressureconcurrentexecutiontests testcoordinatingprimarythreadedupdatetoshardlimitsandrejections shardindexingpressureconcurrentexecutiontests java | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.