Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
65,281 | 19,321,013,201 | IssuesEvent | 2021-12-14 05:34:12 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [🐛 Bug]: Not stable when tested in Edge IE mode with IEDriver 4.0.0.0 | I-defect needs-triaging | ### What happened?
Since previous IEDriver had a problem with opening child windows in Edge IE mode, we have downloaded the new IEDriver 4.0.0.0 and been testing the stablility.
Recursively accessing a page with window.open and calling getWindowHandler through the new IEDriver, selenium access will become unstable in 2 ways.
First, when recursive access reaches about 20 times, it will be unable to get the window handle.
Using the normal IE or normal Edge, the problem with not occur and could get the child window handle permanently, so this problem seems only to occur in IE mode.
Second, selenium timeout occurs.
Exception example)
Exception in thread "main" org.openqa.selenium.TimeoutException: java.util.concurrent.TimeoutException
The timeout occurs in many methods, like close(closing driver), getWindowhandlers, findelement(xpath), possibilty the driver control has a problem.
In closing we want to know if this problem reproduces in every windows environment, and the cause of this problem
[Test.txt](https://github.com/SeleniumHQ/selenium/files/7708951/Test.txt)
[IEMode_1st.txt](https://github.com/SeleniumHQ/selenium/files/7708952/IEMode_1st.txt)[IEMode_2nd.txt](https://github.com/SeleniumHQ/selenium/files/7708947/IEMode_2nd.txt)
[IEMode_3rd.txt](https://github.com/SeleniumHQ/selenium/files/7708948/IEMode_3rd.txt)
[IEMode_4th.txt](https://github.com/SeleniumHQ/selenium/files/7708949/IEMode_4th.txt)
[IEmode_timeout.txt](https://github.com/SeleniumHQ/selenium/files/7708950/IEmode_timeout.txt)
.
### How can we reproduce the issue?
```shell
We have made a simple test program Test.java(Test.txt) that recurisively window.open and get handler of the child window.
When it fails to get the handle, the program will retry 10 times to get the window handle, but in this case no matter you retry you wouldn't be able to get the window handle.
On failing the 10 time retry, it will abort and end the program.
When ending the program it will output which loop it failed to get the handle.
(Otherwise it means it successed n-1 loops to get the handle.)
The result, we have tried about 4 cycles, each cycle was
1st. 38 loops
2nd. 34 loops
3rd. 45 loops
4th. 5 loops
For the timeout ocasion sample, please refer IEmode_timeout.txt
```
### Relevant log output
```shell
Refer to the attached file
IEMode_1st.txt
IEMode_2nd.txt
IEMode_3rd.txt
IEMode_4th.txt
IEmode_timeout.txt
```
### Operating System
Windows 10
### Selenium version
Java 4.0.0
### What are the browser(s) and version(s) where you see this issue?
Edge v.96.0.1054.43 (IE Mode)
### What are the browser driver(s) and version(s) where you see this issue?
Internet Explorer Driver 4.0.0(32bit)
### Are you using Selenium Grid?
No | 1.0 | [🐛 Bug]: Not stable when tested in Edge IE mode with IEDriver 4.0.0.0 - ### What happened?
Since previous IEDriver had a problem with opening child windows in Edge IE mode, we have downloaded the new IEDriver 4.0.0.0 and been testing the stablility.
Recursively accessing a page with window.open and calling getWindowHandler through the new IEDriver, selenium access will become unstable in 2 ways.
First, when recursive access reaches about 20 times, it will be unable to get the window handle.
Using the normal IE or normal Edge, the problem with not occur and could get the child window handle permanently, so this problem seems only to occur in IE mode.
Second, selenium timeout occurs.
Exception example)
Exception in thread "main" org.openqa.selenium.TimeoutException: java.util.concurrent.TimeoutException
The timeout occurs in many methods, like close(closing driver), getWindowhandlers, findelement(xpath), possibilty the driver control has a problem.
In closing we want to know if this problem reproduces in every windows environment, and the cause of this problem
[Test.txt](https://github.com/SeleniumHQ/selenium/files/7708951/Test.txt)
[IEMode_1st.txt](https://github.com/SeleniumHQ/selenium/files/7708952/IEMode_1st.txt)[IEMode_2nd.txt](https://github.com/SeleniumHQ/selenium/files/7708947/IEMode_2nd.txt)
[IEMode_3rd.txt](https://github.com/SeleniumHQ/selenium/files/7708948/IEMode_3rd.txt)
[IEMode_4th.txt](https://github.com/SeleniumHQ/selenium/files/7708949/IEMode_4th.txt)
[IEmode_timeout.txt](https://github.com/SeleniumHQ/selenium/files/7708950/IEmode_timeout.txt)
.
### How can we reproduce the issue?
```shell
We have made a simple test program Test.java(Test.txt) that recurisively window.open and get handler of the child window.
When it fails to get the handle, the program will retry 10 times to get the window handle, but in this case no matter you retry you wouldn't be able to get the window handle.
On failing the 10 time retry, it will abort and end the program.
When ending the program it will output which loop it failed to get the handle.
(Otherwise it means it successed n-1 loops to get the handle.)
The result, we have tried about 4 cycles, each cycle was
1st. 38 loops
2nd. 34 loops
3rd. 45 loops
4th. 5 loops
For the timeout ocasion sample, please refer IEmode_timeout.txt
```
### Relevant log output
```shell
Refer to the attached file
IEMode_1st.txt
IEMode_2nd.txt
IEMode_3rd.txt
IEMode_4th.txt
IEmode_timeout.txt
```
### Operating System
Windows 10
### Selenium version
Java 4.0.0
### What are the browser(s) and version(s) where you see this issue?
Edge v.96.0.1054.43 (IE Mode)
### What are the browser driver(s) and version(s) where you see this issue?
Internet Explorer Driver 4.0.0(32bit)
### Are you using Selenium Grid?
No | defect | not stable when tested in edge ie mode with iedriver what happened since previous iedriver had a problem with opening child windows in edge ie mode we have downloaded the new iedriver and been testing the stablility recursively accessing a page with window open and calling getwindowhandler through the new iedriver selenium access will become unstable in ways first when recursive access reaches about times it will be unable to get the window handle using the normal ie or normal edge the problem with not occur and could get the child window handle permanently so this problem seems only to occur in ie mode second selenium timeout occurs exception example exception in thread main org openqa selenium timeoutexception java util concurrent timeoutexception the timeout occurs in many methods like close closing driver getwindowhandlers findelement xpath possibilty the driver control has a problem in closing we want to know if this problem reproduces in every windows environment and the cause of this problem how can we reproduce the issue shell we have made a simple test program test java test txt that recurisively window open and get handler of the child window when it fails to get the handle the program will retry times to get the window handle but in this case no matter you retry you wouldn t be able to get the window handle on failing the time retry it will abort and end the program when ending the program it will output which loop it failed to get the handle otherwise it means it successed n loops to get the handle the result we have tried about cycles each cycle was loops loops loops loops for the timeout ocasion sample please refer iemode timeout txt relevant log output shell refer to the attached file iemode txt iemode txt iemode txt iemode txt iemode timeout txt operating system windows selenium version java what are the browser s and version s where you see this issue edge v ie mode what are the browser driver s and version s where you see this issue internet explorer driver are you using selenium grid no | 1 |
146,153 | 5,612,132,448 | IssuesEvent | 2017-04-03 03:03:22 | Templarian/MaterialDesign | https://api.github.com/repos/Templarian/MaterialDesign | closed | Rename Network Icons | High Priority | - [x] `network-upload` > `upload-network`
- [x] `network-download` > `download-network`
- [x] `network-question` > `help-network` | 1.0 | Rename Network Icons - - [x] `network-upload` > `upload-network`
- [x] `network-download` > `download-network`
- [x] `network-question` > `help-network` | non_defect | rename network icons network upload upload network network download download network network question help network | 0 |
116,048 | 9,818,968,584 | IssuesEvent | 2019-06-13 20:39:48 | aces/Loris | https://api.github.com/repos/aces/Loris | closed | [Imaging Uploader] - Incorrect behaviour when uploading a huge file | 21.0.0 Testing Bug Fix PR sent | The file size limit for file uploads has been set to 2Mb on the testing VM. I tried uploading a 1Gb file and I got the 'Submission error' popup, but there was no indication of what went wrong. When I clicked on the popup to dismiss it, the popup disappeared but the imaging uploader froze afterwards and the upload tab became unusable. | 1.0 | [Imaging Uploader] - Incorrect behaviour when uploading a huge file - The file size limit for file uploads has been set to 2Mb on the testing VM. I tried uploading a 1Gb file and I got the 'Submission error' popup, but there was no indication of what went wrong. When I clicked on the popup to dismiss it, the popup disappeared but the imaging uploader froze afterwards and the upload tab became unusable. | non_defect | incorrect behaviour when uploading a huge file the file size limit for file uploads has been set to on the testing vm i tried uploading a file and i got the submission error popup but there was no indication of what went wrong when i clicked on the popup to dismiss it the popup disappeared but the imaging uploader froze afterwards and the upload tab became unusable | 0 |
71,641 | 3,366,122,828 | IssuesEvent | 2015-11-21 03:21:26 | hackthesystemATX/doctrinr | https://api.github.com/repos/hackthesystemATX/doctrinr | reopened | User Flow Diagram | high priority | Someone needs to creat a low-fidelity diagram of the userflow.
[Example] (https://www.google.com/search?q=userflow&source=lnms&tbm=isch&sa=X&ved=0ahUKEwjg9KfPxaDJAhWCUogKHTYzC0cQ_AUICCgC&biw=1036&bih=805)
Something simple so we can easily communicate it to a desiger/developer. | 1.0 | User Flow Diagram - Someone needs to creat a low-fidelity diagram of the userflow.
[Example] (https://www.google.com/search?q=userflow&source=lnms&tbm=isch&sa=X&ved=0ahUKEwjg9KfPxaDJAhWCUogKHTYzC0cQ_AUICCgC&biw=1036&bih=805)
Something simple so we can easily communicate it to a desiger/developer. | non_defect | user flow diagram someone needs to creat a low fidelity diagram of the userflow something simple so we can easily communicate it to a desiger developer | 0 |
47,713 | 7,344,447,819 | IssuesEvent | 2018-03-07 14:42:59 | rhpit/manageiq-cli | https://api.github.com/repos/rhpit/manageiq-cli | closed | Document how to add/delete OpenStack provider | documentation | ### Summary
To document how to add and delete OpenStack providers. | 1.0 | Document how to add/delete OpenStack provider - ### Summary
To document how to add and delete OpenStack providers. | non_defect | document how to add delete openstack provider summary to document how to add and delete openstack providers | 0 |
53,940 | 13,262,537,900 | IssuesEvent | 2020-08-20 22:00:35 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [dataclasses] pretty printing for I3RecoPulseSeriesMapMask (Trac #2365) | Migrated from Trac combo core defect | I attached some screenshots to explain. It had already disappeared in icerec 5.2.3. But, I really like to have it back.
To demonstrate open an L2 file
`/data/user/serap/icerec518/build/env-shell.sh dataio-pyshovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`
goto first Q frame, press return on any I3RecoPulseSeriesMapMask, you’see ll the contents in human readable format, press x you’ll see the content in xml format.
now check it out the same with combo
`/data/user/serap/combo-trunk/build/env-shell.sh dataio-shovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`
press return on InIceDSTOnlyPulses, you’ll see a bitmap or something.
How can we bring the good old human readable format back?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2365">https://code.icecube.wisc.edu/projects/icecube/ticket/2365</a>, reported by tilavand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"_ts": "1593001902142004",
"description": "I attached some screenshots to explain. It had already disappeared in icerec 5.2.3. But, I really like to have it back.\n\nTo demonstrate open an L2 file\n`/data/user/serap/icerec518/build/env-shell.sh dataio-pyshovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`\n\ngoto first Q frame, press return on any I3RecoPulseSeriesMapMask, you\u2019see ll the contents in human readable format, press x you\u2019ll see the content in xml format.\n\nnow check it out the same with combo\n`/data/user/serap/combo-trunk/build/env-shell.sh dataio-shovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`\npress return on InIceDSTOnlyPulses, you\u2019ll see a bitmap or something.\n\nHow can we bring the good old human readable format back?",
"reporter": "tilav",
"cc": "david.schultz, olivas",
"resolution": "fixed",
"time": "2019-10-13T16:43:35",
"component": "combo core",
"summary": "[dataclasses] pretty printing for I3RecoPulseSeriesMapMask",
"priority": "normal",
"keywords": "",
"milestone": "Autumnal Equinox 2020",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [dataclasses] pretty printing for I3RecoPulseSeriesMapMask (Trac #2365) - I attached some screenshots to explain. It had already disappeared in icerec 5.2.3. But, I really like to have it back.
To demonstrate open an L2 file
`/data/user/serap/icerec518/build/env-shell.sh dataio-pyshovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`
goto first Q frame, press return on any I3RecoPulseSeriesMapMask, you’see ll the contents in human readable format, press x you’ll see the content in xml format.
now check it out the same with combo
`/data/user/serap/combo-trunk/build/env-shell.sh dataio-shovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`
press return on InIceDSTOnlyPulses, you’ll see a bitmap or something.
How can we bring the good old human readable format back?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2365">https://code.icecube.wisc.edu/projects/icecube/ticket/2365</a>, reported by tilavand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"_ts": "1593001902142004",
"description": "I attached some screenshots to explain. It had already disappeared in icerec 5.2.3. But, I really like to have it back.\n\nTo demonstrate open an L2 file\n`/data/user/serap/icerec518/build/env-shell.sh dataio-pyshovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`\n\ngoto first Q frame, press return on any I3RecoPulseSeriesMapMask, you\u2019see ll the contents in human readable format, press x you\u2019ll see the content in xml format.\n\nnow check it out the same with combo\n`/data/user/serap/combo-trunk/build/env-shell.sh dataio-shovel /data/exp/IceCube/2019/filtered/level2/0929/Run00133112/Level2_IC86.2019_data_Run00133112_Subrun00000000_00000216.i3.zst`\npress return on InIceDSTOnlyPulses, you\u2019ll see a bitmap or something.\n\nHow can we bring the good old human readable format back?",
"reporter": "tilav",
"cc": "david.schultz, olivas",
"resolution": "fixed",
"time": "2019-10-13T16:43:35",
"component": "combo core",
"summary": "[dataclasses] pretty printing for I3RecoPulseSeriesMapMask",
"priority": "normal",
"keywords": "",
"milestone": "Autumnal Equinox 2020",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| defect | pretty printing for trac i attached some screenshots to explain it had already disappeared in icerec but i really like to have it back to demonstrate open an file data user serap build env shell sh dataio pyshovel data exp icecube filtered data zst goto first q frame press return on any you’see ll the contents in human readable format press x you’ll see the content in xml format now check it out the same with combo data user serap combo trunk build env shell sh dataio shovel data exp icecube filtered data zst press return on inicedstonlypulses you’ll see a bitmap or something how can we bring the good old human readable format back migrated from json status closed changetime ts description i attached some screenshots to explain it had already disappeared in icerec but i really like to have it back n nto demonstrate open an file n data user serap build env shell sh dataio pyshovel data exp icecube filtered data zst n ngoto first q frame press return on any you ll the contents in human readable format press x you see the content in xml format n nnow check it out the same with combo n data user serap combo trunk build env shell sh dataio shovel data exp icecube filtered data zst npress return on inicedstonlypulses you see a bitmap or something n nhow can we bring the good old human readable format back reporter tilav cc david schultz olivas resolution fixed time component combo core summary pretty printing for priority normal keywords milestone autumnal equinox owner olivas type defect | 1 |
164,518 | 25,981,239,500 | IssuesEvent | 2022-12-19 19:03:17 | HackIllinois/iOS | https://api.github.com/repos/HackIllinois/iOS | opened | Update Event Card | enhancement feature design | - Create label for event type
- Company (for only applicable events)
- Location
- Map in popup
- Update UI: corner radius, colors, fonts
- Keep remaining features | 1.0 | Update Event Card - - Create label for event type
- Company (for only applicable events)
- Location
- Map in popup
- Update UI: corner radius, colors, fonts
- Keep remaining features | non_defect | update event card create label for event type company for only applicable events location map in popup update ui corner radius colors fonts keep remaining features | 0 |
44,896 | 23,811,618,051 | IssuesEvent | 2022-09-04 21:00:39 | JuliaRobotics/DistributedFactorGraphs.jl | https://api.github.com/repos/JuliaRobotics/DistributedFactorGraphs.jl | closed | createDfgSessionIfNotExist should be optimized | performance | This call makes between 3-6 Neo4j calls, which could be made into a single call. | True | createDfgSessionIfNotExist should be optimized - This call makes between 3-6 Neo4j calls, which could be made into a single call. | non_defect | createdfgsessionifnotexist should be optimized this call makes between calls which could be made into a single call | 0 |
46,354 | 7,253,240,967 | IssuesEvent | 2018-02-16 03:53:51 | numpy/numpy | https://api.github.com/repos/numpy/numpy | closed | The circleci test is broken. | 00 - Bug 03 - Documentation | I suspect this is related to the recent release of sphinx 1.7.0 on 2/12/2018. We may need to update the submodule hash for sphinxext, or maybe even fix numpydoc. | 1.0 | The circleci test is broken. - I suspect this is related to the recent release of sphinx 1.7.0 on 2/12/2018. We may need to update the submodule hash for sphinxext, or maybe even fix numpydoc. | non_defect | the circleci test is broken i suspect this is related to the recent release of sphinx on we may need to update the submodule hash for sphinxext or maybe even fix numpydoc | 0 |
68,050 | 17,138,060,250 | IssuesEvent | 2021-07-13 06:15:04 | tsunamayo/Starship-EVO | https://api.github.com/repos/tsunamayo/Starship-EVO | opened | [New build - EXPERIMENTAL] 21w28a: Docking Port and sub-modules | Build Release Note | This build introduce the new docking port and a new Docking Controller allowing you to spawn decorative prefabs in a random fashion.
Check out the upcoming post to learn more about the new docking philosophy.
Docking New Features
- New Docking Connector brick. It can grid-resize.
- New Docking Controller brick.
- Configure "spawning sub-modules" inside the controller.
- Add tags to be spawned with a given weight for the randomized draw.
- Any entity with a docking module can be used as a sub-module, to be configured inside the codex entity windows.
Still to do:
- Connector size check for spawn
- Station Module docking uses new docking port.
- Kiosks spawning for stations
- Spaceship universal docking port.
Also new due to popular request:
- Chain link: simply holding Left-SHIFT during a link action will hold-on to the current link.
- Box delete tool: select a volume like with the copy/paste. Then right click to delete the selection.
- F3 Option: disable NPC spawn
- F3 Option: peaceful NPCs mode.
Bugfixes:
#3987 Station Starting block name. | 1.0 | [New build - EXPERIMENTAL] 21w28a: Docking Port and sub-modules - This build introduce the new docking port and a new Docking Controller allowing you to spawn decorative prefabs in a random fashion.
Check out the upcoming post to learn more about the new docking philosophy.
Docking New Features
- New Docking Connector brick. It can grid-resize.
- New Docking Controller brick.
- Configure "spawning sub-modules" inside the controller.
- Add tags to be spawned with a given weight for the randomized draw.
- Any entity with a docking module can be used as a sub-module, to be configured inside the codex entity windows.
Still to do:
- Connector size check for spawn
- Station Module docking uses new docking port.
- Kiosks spawning for stations
- Spaceship universal docking port.
Also new due to popular request:
- Chain link: simply holding Left-SHIFT during a link action will hold-on to the current link.
- Box delete tool: select a volume like with the copy/paste. Then right click to delete the selection.
- F3 Option: disable NPC spawn
- F3 Option: peaceful NPCs mode.
Bugfixes:
#3987 Station Starting block name. | non_defect | docking port and sub modules this build introduce the new docking port and a new docking controller allowing you to spawn decorative prefabs in a random fashion check out the upcoming post to learn more about the new docking philosophy docking new features new docking connector brick it can grid resize new docking controller brick configure spawning sub modules inside the controller add tags to be spawned with a given weight for the randomized draw any entity with a docking module can be used as a sub module to be configured inside the codex entity windows still to do connector size check for spawn station module docking uses new docking port kiosks spawning for stations spaceship universal docking port also new due to popular request chain link simply holding left shift during a link action will hold on to the current link box delete tool select a volume like with the copy paste then right click to delete the selection option disable npc spawn option peaceful npcs mode bugfixes station starting block name | 0 |
53,136 | 13,260,998,833 | IssuesEvent | 2020-08-20 19:12:33 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | Stagers should set scratch directory lazily (Trac #790) | Migrated from Trac combo core defect | The I3FileStagers need a scratch directory to store temporary files. Currently this must be set in the constructor, and is guessed from certain commonly used environment variables. If none of the usual suspects are set, it just fails.
It would be better to move this to a base class function, and allow the user to set a global scratch directory for the process programatically.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/790">https://code.icecube.wisc.edu/projects/icecube/ticket/790</a>, reported by jvansanten</summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-21T23:53:53",
"_ts": "1413935633157956",
"description": "The I3FileStagers need a scratch directory to store temporary files. Currently this must be set in the constructor, and is guessed from certain commonly used environment variables. If none of the usual suspects are set, it just fails.\n\nIt would be better to move this to a base class function, and allow the user to set a global scratch directory for the process programatically. ",
"reporter": "jvansanten",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"time": "2014-10-21T20:02:53",
"component": "combo core",
"summary": "Stagers should set scratch directory lazily",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Stagers should set scratch directory lazily (Trac #790) - The I3FileStagers need a scratch directory to store temporary files. Currently this must be set in the constructor, and is guessed from certain commonly used environment variables. If none of the usual suspects are set, it just fails.
It would be better to move this to a base class function, and allow the user to set a global scratch directory for the process programatically.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/790">https://code.icecube.wisc.edu/projects/icecube/ticket/790</a>, reported by jvansanten</summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-21T23:53:53",
"_ts": "1413935633157956",
"description": "The I3FileStagers need a scratch directory to store temporary files. Currently this must be set in the constructor, and is guessed from certain commonly used environment variables. If none of the usual suspects are set, it just fails.\n\nIt would be better to move this to a base class function, and allow the user to set a global scratch directory for the process programatically. ",
"reporter": "jvansanten",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"time": "2014-10-21T20:02:53",
"component": "combo core",
"summary": "Stagers should set scratch directory lazily",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | stagers should set scratch directory lazily trac the need a scratch directory to store temporary files currently this must be set in the constructor and is guessed from certain commonly used environment variables if none of the usual suspects are set it just fails it would be better to move this to a base class function and allow the user to set a global scratch directory for the process programatically migrated from json status closed changetime ts description the need a scratch directory to store temporary files currently this must be set in the constructor and is guessed from certain commonly used environment variables if none of the usual suspects are set it just fails n nit would be better to move this to a base class function and allow the user to set a global scratch directory for the process programatically reporter jvansanten cc david schultz icecube wisc edu resolution fixed time component combo core summary stagers should set scratch directory lazily priority normal keywords milestone owner type defect | 1 |
265,380 | 23,163,769,705 | IssuesEvent | 2022-07-29 21:05:46 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: restore2TB/nodes=6/cpus=8/pd-volume=2500GB failed | C-test-failure O-robot O-roachtest branch-master release-blocker T-bulkio | roachtest.restore2TB/nodes=6/cpus=8/pd-volume=2500GB [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5889604?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5889604?buildTab=artifacts#/restore2TB/nodes=6/cpus=8/pd-volume=2500GB) on master @ [7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732](https://github.com/cockroachdb/cockroach/commits/7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732):
```
Wraps: (2) output in run_125413.930258424_n1_cockroach_sql
Wraps: (3) ./cockroach sql --insecure -e "
| RESTORE csv.bank FROM
| 'gs://cockroach-fixtures/workload/bank/version=1.0.0,payload-bytes=10240,ranges=0,rows=65104166,seed=1/bank?AUTH=implicit'
| WITH into_db = 'restore2tb'" returned
| stderr:
| ERROR: importing 21888 ranges: Get "https://storage.googleapis.com/cockroach-fixtures/workload/bank/version=1.0.0,payload-bytes=10240,ranges=0,rows=65104166,seed=1/bank/2-85.sst": stream error: stream ID 7; INTERNAL_ERROR; received from peer
| Failed running "sql"
|
| stdout:
Wraps: (4) COMMAND_PROBLEM
Wraps: (5) Node 1. Command with error:
| ``````
| ./cockroach sql --insecure -e "
| RESTORE csv.bank FROM
| 'gs://cockroach-fixtures/workload/bank/version=1.0.0,payload-bytes=10240,ranges=0,rows=65104166,seed=1/bank?AUTH=implicit'
| WITH into_db = 'restore2tb'"
| ``````
Wraps: (6) exit status 1
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError
monitor.go:127,restore.go:453,test_runner.go:896: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitorImpl).WaitE
| main/pkg/cmd/roachtest/monitor.go:115
| main.(*monitorImpl).Wait
| main/pkg/cmd/roachtest/monitor.go:123
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerRestore.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/restore.go:453
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:896
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitorImpl).wait.func2
| main/pkg/cmd/roachtest/monitor.go:171
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| main/pkg/cmd/roachtest/monitor.go:80
| runtime.doInit
| GOROOT/src/runtime/proc.go:6222
| runtime.main
| GOROOT/src/runtime/proc.go:233
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1571
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=8</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #85216 roachtest: restore2TB/nodes=6/cpus=8/pd-volume=2500GB failed [C-test-failure O-roachtest O-robot T-bulkio branch-release-22.1 release-blocker]
</p>
</details>
/cc @cockroachdb/bulk-io
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore2TB/nodes=6/cpus=8/pd-volume=2500GB.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-18111 | 2.0 | roachtest: restore2TB/nodes=6/cpus=8/pd-volume=2500GB failed - roachtest.restore2TB/nodes=6/cpus=8/pd-volume=2500GB [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5889604?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5889604?buildTab=artifacts#/restore2TB/nodes=6/cpus=8/pd-volume=2500GB) on master @ [7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732](https://github.com/cockroachdb/cockroach/commits/7c95ccd0a9f0b0aca9cdb0ed813caf8d60b28732):
```
Wraps: (2) output in run_125413.930258424_n1_cockroach_sql
Wraps: (3) ./cockroach sql --insecure -e "
| RESTORE csv.bank FROM
| 'gs://cockroach-fixtures/workload/bank/version=1.0.0,payload-bytes=10240,ranges=0,rows=65104166,seed=1/bank?AUTH=implicit'
| WITH into_db = 'restore2tb'" returned
| stderr:
| ERROR: importing 21888 ranges: Get "https://storage.googleapis.com/cockroach-fixtures/workload/bank/version=1.0.0,payload-bytes=10240,ranges=0,rows=65104166,seed=1/bank/2-85.sst": stream error: stream ID 7; INTERNAL_ERROR; received from peer
| Failed running "sql"
|
| stdout:
Wraps: (4) COMMAND_PROBLEM
Wraps: (5) Node 1. Command with error:
| ``````
| ./cockroach sql --insecure -e "
| RESTORE csv.bank FROM
| 'gs://cockroach-fixtures/workload/bank/version=1.0.0,payload-bytes=10240,ranges=0,rows=65104166,seed=1/bank?AUTH=implicit'
| WITH into_db = 'restore2tb'"
| ``````
Wraps: (6) exit status 1
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError
monitor.go:127,restore.go:453,test_runner.go:896: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitorImpl).WaitE
| main/pkg/cmd/roachtest/monitor.go:115
| main.(*monitorImpl).Wait
| main/pkg/cmd/roachtest/monitor.go:123
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerRestore.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/restore.go:453
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:896
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitorImpl).wait.func2
| main/pkg/cmd/roachtest/monitor.go:171
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| main/pkg/cmd/roachtest/monitor.go:80
| runtime.doInit
| GOROOT/src/runtime/proc.go:6222
| runtime.main
| GOROOT/src/runtime/proc.go:233
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1571
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=8</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #85216 roachtest: restore2TB/nodes=6/cpus=8/pd-volume=2500GB failed [C-test-failure O-roachtest O-robot T-bulkio branch-release-22.1 release-blocker]
</p>
</details>
/cc @cockroachdb/bulk-io
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore2TB/nodes=6/cpus=8/pd-volume=2500GB.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-18111 | non_defect | roachtest nodes cpus pd volume failed roachtest nodes cpus pd volume with on master wraps output in run cockroach sql wraps cockroach sql insecure e restore csv bank from gs cockroach fixtures workload bank version payload bytes ranges rows seed bank auth implicit with into db returned stderr error importing ranges get stream error stream id internal error received from peer failed running sql stdout wraps command problem wraps node command with error cockroach sql insecure e restore csv bank from gs cockroach fixtures workload bank version payload bytes ranges rows seed bank auth implicit with into db wraps exit status error types withstack withstack errutil withprefix cluster withcommanddetails errors cmd hintdetail withdetail exec exiterror monitor go restore go test runner go monitor failure monitor task failed t fatal was called attached stack trace stack trace main monitorimpl waite main pkg cmd roachtest monitor go main monitorimpl wait main pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests registerrestore github com cockroachdb cockroach pkg cmd roachtest tests restore go main testrunner runtest main pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace stack trace main monitorimpl wait main pkg cmd roachtest monitor go wraps monitor task failed wraps attached stack trace stack trace main init main pkg cmd roachtest monitor go runtime doinit goroot src runtime proc go runtime main goroot src runtime proc go runtime goexit goroot src runtime asm s wraps t fatal was called error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil leaferror parameters roachtest cloud gce roachtest cpu roachtest ssd help see see same failure on other branches roachtest nodes cpus pd volume failed cc cockroachdb bulk io jira issue crdb | 0 |
62,451 | 17,023,925,644 | IssuesEvent | 2021-07-03 04:35:32 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | name=* is not longer rendered for amenity=school and amenity=kindergarten | Component: mapnik Priority: minor Resolution: duplicate Type: defect | **[Submitted to the original trac issue database at 7.46pm, Saturday, 19th September 2015]**
According to the wiki amenity=school, amenity=kindergarten, amenity=college, amenity=university should be mapped as an area with name=*. The name should not be placed at the buildings. Since some time this dosn't work anymore: The name isn't rendered on any zoom level.
Perhaps a similar style as for names on landuse=industrial should be used?
Example: [http://www.openstreetmap.org/#map=18/47.32913/8.78195] contains a wastwater plant and two schools with a name. | 1.0 | name=* is not longer rendered for amenity=school and amenity=kindergarten - **[Submitted to the original trac issue database at 7.46pm, Saturday, 19th September 2015]**
According to the wiki amenity=school, amenity=kindergarten, amenity=college, amenity=university should be mapped as an area with name=*. The name should not be placed at the buildings. Since some time this dosn't work anymore: The name isn't rendered on any zoom level.
Perhaps a similar style as for names on landuse=industrial should be used?
Example: [http://www.openstreetmap.org/#map=18/47.32913/8.78195] contains a wastwater plant and two schools with a name. | defect | name is not longer rendered for amenity school and amenity kindergarten according to the wiki amenity school amenity kindergarten amenity college amenity university should be mapped as an area with name the name should not be placed at the buildings since some time this dosn t work anymore the name isn t rendered on any zoom level perhaps a similar style as for names on landuse industrial should be used example contains a wastwater plant and two schools with a name | 1 |
182,389 | 21,664,514,097 | IssuesEvent | 2022-05-07 01:36:39 | tt9133github/ForTest2 | https://api.github.com/repos/tt9133github/ForTest2 | closed | CVE-2020-10738 (High) detected in moodle/moodle-v2.6.2 - autoclosed | security vulnerability | ## CVE-2020-10738 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moodle/moodle-v2.6.2</b></p></summary>
<p>Moodle - the world's open source learning platform</p>
<p>
Dependency Hierarchy:
- :x: **moodle/moodle-v2.6.2** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tt9133github/ForTest2/commit/27e1f740a162e55490d9ca929f243afbaa92476c">27e1f740a162e55490d9ca929f243afbaa92476c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in Moodle versions 3.8 before 3.8.3, 3.7 before 3.7.6, 3.6 before 3.6.10, 3.5 before 3.5.12 and earlier unsupported versions. It was possible to create a SCORM package in such a way that when added to a course, it could be interacted with via web services in order to achieve remote code execution.
<p>Publish Date: 2020-05-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10738>CVE-2020-10738</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10738">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10738</a></p>
<p>Release Date: 2020-05-21</p>
<p>Fix Resolution: 3.5.12,3.6.10,3.7.6,3.8.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-10738 (High) detected in moodle/moodle-v2.6.2 - autoclosed - ## CVE-2020-10738 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moodle/moodle-v2.6.2</b></p></summary>
<p>Moodle - the world's open source learning platform</p>
<p>
Dependency Hierarchy:
- :x: **moodle/moodle-v2.6.2** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tt9133github/ForTest2/commit/27e1f740a162e55490d9ca929f243afbaa92476c">27e1f740a162e55490d9ca929f243afbaa92476c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in Moodle versions 3.8 before 3.8.3, 3.7 before 3.7.6, 3.6 before 3.6.10, 3.5 before 3.5.12 and earlier unsupported versions. It was possible to create a SCORM package in such a way that when added to a course, it could be interacted with via web services in order to achieve remote code execution.
<p>Publish Date: 2020-05-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10738>CVE-2020-10738</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10738">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10738</a></p>
<p>Release Date: 2020-05-21</p>
<p>Fix Resolution: 3.5.12,3.6.10,3.7.6,3.8.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in moodle moodle autoclosed cve high severity vulnerability vulnerable library moodle moodle moodle the world s open source learning platform dependency hierarchy x moodle moodle vulnerable library found in head commit a href vulnerability details a flaw was found in moodle versions before before before before and earlier unsupported versions it was possible to create a scorm package in such a way that when added to a course it could be interacted with via web services in order to achieve remote code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
60,624 | 17,023,475,179 | IssuesEvent | 2021-07-03 02:13:16 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | name of waterway not always completely visible | Component: mapnik Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 8.55pm, Friday, 4th September 2009]**
It seems that in some cases the name of a waterway is not completely visible because a road is drawn on top of it.
There are 2 examples here: http://www.openstreetmap.org/?lat=50.9235&lon=4.65383&zoom=15&layers=B000FTF
It's probably both times drawn in the middle of the piece of waterway not aware that something else is being drawn there on a higher layer.
| 1.0 | name of waterway not always completely visible - **[Submitted to the original trac issue database at 8.55pm, Friday, 4th September 2009]**
It seems that in some cases the name of a waterway is not completely visible because a road is drawn on top of it.
There are 2 examples here: http://www.openstreetmap.org/?lat=50.9235&lon=4.65383&zoom=15&layers=B000FTF
It's probably both times drawn in the middle of the piece of waterway not aware that something else is being drawn there on a higher layer.
| defect | name of waterway not always completely visible it seems that in some cases the name of a waterway is not completely visible because a road is drawn on top of it there are examples here it s probably both times drawn in the middle of the piece of waterway not aware that something else is being drawn there on a higher layer | 1 |
67,054 | 20,825,586,286 | IssuesEvent | 2022-03-18 20:24:30 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | Energy Conservation for STFT | defect scipy.signal | The project I am working on requires a short-time Fourier transformation. For this, I use scipy.signal.stft. From theory, we learn that the signal in time domain has the same energy as in time-frequency domain; Parseval's theorem (https://en.wikipedia.org/wiki/Parseval%27s_theorem). However, when I use scipy.signal.stft, I do not obtain the same amount of energy if I compare the computation for both values of energy. I am pretty sure, it is because there is some scaling involved when using scipy.signal.stft, so I spit through the source code. After some deep search, I have found a scaling factor of np.sqrt(1/(sum(win)**2)). However, using this scaling factor, does not work:
Example code:
```
fs = 10e3
N = 1e5
amp = 2 * np.sqrt(2)
noise_power = 0.01 * fs / 2
time = np.arange(N) / float(fs)
mod = 500*np.cos(2*np.pi*0.25*time)
carrier = amp * np.sin(2*np.pi*3e3*time + mod)
noise = np.random.normal(scale=np.sqrt(noise_power),
size=time.shape)
noise *= np.exp(-time/5)
x = carrier + noise
frame_size = 1000
window = scipy.signal.windows.hann(frame_size,sym=False)
f, t, Zxx = scipy.signal.stft(x, fs, window=window, nperseg=frame_size)
scale = np.sqrt(1/(sum(window)**2))
Zxx_scaled = Zxx/scale
Et=np.sum(x**2)
Ef=np.sum(abs(Zxx_scaled)**2)/N
print("Et = %.3f" % Et)
print( "Ef = %.3f" % Ef)
```
results in (depending on noise, but in the scale of):
```
Et = 1641328.607
Ef = 6158.539
```
Does anyone know how to properly scale the resulting STFT matrix values to obtain equal energy levels and prove energy conservation for the STFT?
(I am looking for a `sum(window)**2` constant transformation (and window) and, as the window is used in the STFT and I-STFT, I assume I have a 'tight frame' transformation if I use the hann window. However, proving energy conservation is a first step.....)
| 1.0 | Energy Conservation for STFT - The project I am working on requires a short-time Fourier transformation. For this, I use scipy.signal.stft. From theory, we learn that the signal in time domain has the same energy as in time-frequency domain; Parseval's theorem (https://en.wikipedia.org/wiki/Parseval%27s_theorem). However, when I use scipy.signal.stft, I do not obtain the same amount of energy if I compare the computation for both values of energy. I am pretty sure, it is because there is some scaling involved when using scipy.signal.stft, so I spit through the source code. After some deep search, I have found a scaling factor of np.sqrt(1/(sum(win)**2)). However, using this scaling factor, does not work:
Example code:
```
fs = 10e3
N = 1e5
amp = 2 * np.sqrt(2)
noise_power = 0.01 * fs / 2
time = np.arange(N) / float(fs)
mod = 500*np.cos(2*np.pi*0.25*time)
carrier = amp * np.sin(2*np.pi*3e3*time + mod)
noise = np.random.normal(scale=np.sqrt(noise_power),
size=time.shape)
noise *= np.exp(-time/5)
x = carrier + noise
frame_size = 1000
window = scipy.signal.windows.hann(frame_size,sym=False)
f, t, Zxx = scipy.signal.stft(x, fs, window=window, nperseg=frame_size)
scale = np.sqrt(1/(sum(window)**2))
Zxx_scaled = Zxx/scale
Et=np.sum(x**2)
Ef=np.sum(abs(Zxx_scaled)**2)/N
print("Et = %.3f" % Et)
print( "Ef = %.3f" % Ef)
```
results in (depending on noise, but in the scale of):
```
Et = 1641328.607
Ef = 6158.539
```
Does anyone know how to properly scale the resulting STFT matrix values to obtain equal energy levels and prove energy conservation for the STFT?
(I am looking for a `sum(window)**2` constant transformation (and window) and, as the window is used in the STFT and I-STFT, I assume I have a 'tight frame' transformation if I use the hann window. However, proving energy conservation is a first step.....)
| defect | energy conservation for stft the project i am working on requires a short time fourier transformation for this i use scipy signal stft from theory we learn that the signal in time domain has the same energy as in time frequency domain parseval s theorem however when i use scipy signal stft i do not obtain the same amount of energy if i compare the computation for both values of energy i am pretty sure it is because there is some scaling involved when using scipy signal stft so i spit through the source code after some deep search i have found a scaling factor of np sqrt sum win however using this scaling factor does not work example code fs n amp np sqrt noise power fs time np arange n float fs mod np cos np pi time carrier amp np sin np pi time mod noise np random normal scale np sqrt noise power size time shape noise np exp time x carrier noise frame size window scipy signal windows hann frame size sym false f t zxx scipy signal stft x fs window window nperseg frame size scale np sqrt sum window zxx scaled zxx scale et np sum x ef np sum abs zxx scaled n print et et print ef ef results in depending on noise but in the scale of et ef does anyone know how to properly scale the resulting stft matrix values to obtain equal energy levels and prove energy conservation for the stft i am looking for a sum window constant transformation and window and as the window is used in the stft and i stft i assume i have a tight frame transformation if i use the hann window however proving energy conservation is a first step | 1 |
381,300 | 11,275,905,297 | IssuesEvent | 2020-01-14 21:50:28 | googleapis/java-phishingprotection | https://api.github.com/repos/googleapis/java-phishingprotection | closed | Synthesis failed for java-phishingprotection | autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate java-phishingprotection. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/482a7766-2134-4c8a-887d-f6441089f633).
| 1.0 | Synthesis failed for java-phishingprotection - Hello! Autosynth couldn't regenerate java-phishingprotection. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/482a7766-2134-4c8a-887d-f6441089f633).
| non_defect | synthesis failed for java phishingprotection hello autosynth couldn t regenerate java phishingprotection broken heart here s the output from running synth py cloning into working repo switched to branch autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log | 0 |
65,716 | 19,666,722,643 | IssuesEvent | 2022-01-10 23:43:05 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | closed | Crash on custom icon map with heavylog version | Type-Defect Priority-High | To reproduce the crash:
1) Install this map: [map12121.zip](https://github.com/dkfans/keeperfx/files/7798408/map12121.zip)
2) Run it with keeperfx_hvlog.exe
-> crash straight on load
The map is special for having custom icons and custom sprites.
Found in r2557
| 1.0 | Crash on custom icon map with heavylog version - To reproduce the crash:
1) Install this map: [map12121.zip](https://github.com/dkfans/keeperfx/files/7798408/map12121.zip)
2) Run it with keeperfx_hvlog.exe
-> crash straight on load
The map is special for having custom icons and custom sprites.
Found in r2557
| defect | crash on custom icon map with heavylog version to reproduce the crash install this map run it with keeperfx hvlog exe crash straight on load the map is special for having custom icons and custom sprites found in | 1 |
38,536 | 8,880,735,237 | IssuesEvent | 2019-01-14 07:40:39 | line/armeria | https://api.github.com/repos/line/armeria | closed | Do not log 'Sent a GOAWAY frame: debugData="Error flushing"' | defect | .. because it basically means the connection is already over or the connection has been reset.
Related: #1401
| 1.0 | Do not log 'Sent a GOAWAY frame: debugData="Error flushing"' - .. because it basically means the connection is already over or the connection has been reset.
Related: #1401
| defect | do not log sent a goaway frame debugdata error flushing because it basically means the connection is already over or the connection has been reset related | 1 |
69,589 | 22,551,870,148 | IssuesEvent | 2022-06-27 06:38:52 | scipy/scipy | https://api.github.com/repos/scipy/scipy | reopened | DOC: stats: documentation is not distribution-specific | defect scipy.stats Documentation wontfix | I think that the docstring of the scipy.stat.mean.cdf "the cumulative distribution function of the mean distribution." doesnot match with the scipy documentation. The [documentation page](https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.norm.html#scipy.stats.norm) of the `scipy.stats.norm.cdf` indicates that the second argument is
location. Yet, the docstring saying that it is the 3rd argument.
```
import scipy
scipy.stats.norm.cdf?
```
### The docstring of the scipt.stats.mean.cdf
```
Signature: scipy.stats.norm.cdf(x, *args, **kwds)
Docstring:
Cumulative distribution function of the given RV.
Parameters
x : array_like
quantiles
arg1, arg2, arg3,... : array_like
The shape parameter(s) for the distribution (see docstring of the
instance object for more information)
loc : array_like, optional
location parameter (default=0)
scale : array_like, optional
scale parameter (default=1)
Returns
cdf : ndarray
Cumulative distribution function evaluated at `x`
File: ~/anaconda3/lib/python3.7/site-packages/scipy/stats/_distn_infrastructure.py
Type: method
```
### Scipy/Numpy/Python version information:
1.2.1 1.16.2 sys.version_info(major=3, minor=7, micro=3, releaselevel='final', serial=0) | 1.0 | DOC: stats: documentation is not distribution-specific - I think that the docstring of the scipy.stat.mean.cdf "the cumulative distribution function of the mean distribution." doesnot match with the scipy documentation. The [documentation page](https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.norm.html#scipy.stats.norm) of the `scipy.stats.norm.cdf` indicates that the second argument is
location. Yet, the docstring saying that it is the 3rd argument.
```
import scipy
scipy.stats.norm.cdf?
```
### The docstring of the scipt.stats.mean.cdf
```
Signature: scipy.stats.norm.cdf(x, *args, **kwds)
Docstring:
Cumulative distribution function of the given RV.
Parameters
x : array_like
quantiles
arg1, arg2, arg3,... : array_like
The shape parameter(s) for the distribution (see docstring of the
instance object for more information)
loc : array_like, optional
location parameter (default=0)
scale : array_like, optional
scale parameter (default=1)
Returns
cdf : ndarray
Cumulative distribution function evaluated at `x`
File: ~/anaconda3/lib/python3.7/site-packages/scipy/stats/_distn_infrastructure.py
Type: method
```
### Scipy/Numpy/Python version information:
1.2.1 1.16.2 sys.version_info(major=3, minor=7, micro=3, releaselevel='final', serial=0) | defect | doc stats documentation is not distribution specific i think that the docstring of the scipy stat mean cdf the cumulative distribution function of the mean distribution doesnot match with the scipy documentation the of the scipy stats norm cdf indicates that the second argument is location yet the docstring saying that it is the argument import scipy scipy stats norm cdf the docstring of the scipt stats mean cdf signature scipy stats norm cdf x args kwds docstring cumulative distribution function of the given rv parameters x array like quantiles array like the shape parameter s for the distribution see docstring of the instance object for more information loc array like optional location parameter default scale array like optional scale parameter default returns cdf ndarray cumulative distribution function evaluated at x file lib site packages scipy stats distn infrastructure py type method scipy numpy python version information sys version info major minor micro releaselevel final serial | 1 |
264,772 | 23,138,010,895 | IssuesEvent | 2022-07-28 15:45:41 | nicolargo/glances | https://api.github.com/repos/nicolargo/glances | closed | Config to disable all plugins by default (or enable an exclusive list | enhancement needs test | Thank you for making glances! It's awesome, and I've been looking into using it in combination with Home Assistant to monitor system stats on my network.
**Is your feature request related to a problem? Please describe.**
When deploying glances as a JSON API only, I'd like to be able to limit the stats exposed -- monitoring simple stats like CPU and memory are very useful to me, but I would prefer to not expose things like the process list and software versions to the network, since they could aid automated vulnerability finders.
**Describe the solution you'd like**
It'd be really useful to either be able to disable all plugins by default. For example, support for an argument like `--disable-plugin=*` or `--disable-plugin=all` would work well for this. Alternatively, it would be useful to be able to specify a list of plugins like `--plugins=cpu,mem` and have those be the only plugins loaded.
**Describe alternatives you've considered**
Currently, the only way I've found to accomplish this is listing the entire set of available plugins via:
```
--disable-plugin=alert,amps,cloud,connections,core,cpu,diskio,docker,folders,fs,gpu,help,ip,irq,load,mem,memswap,network,now,percpu,ports,processcount,processlist,psutilversion,quicklook,raid,sensors,smart,system,uptime,wifi
```
This isn't a total solution, because it's possible a future glances update will introduce a new plugin which is default-on. | 1.0 | Config to disable all plugins by default (or enable an exclusive list - Thank you for making glances! It's awesome, and I've been looking into using it in combination with Home Assistant to monitor system stats on my network.
**Is your feature request related to a problem? Please describe.**
When deploying glances as a JSON API only, I'd like to be able to limit the stats exposed -- monitoring simple stats like CPU and memory are very useful to me, but I would prefer to not expose things like the process list and software versions to the network, since they could aid automated vulnerability finders.
**Describe the solution you'd like**
It'd be really useful to either be able to disable all plugins by default. For example, support for an argument like `--disable-plugin=*` or `--disable-plugin=all` would work well for this. Alternatively, it would be useful to be able to specify a list of plugins like `--plugins=cpu,mem` and have those be the only plugins loaded.
**Describe alternatives you've considered**
Currently, the only way I've found to accomplish this is listing the entire set of available plugins via:
```
--disable-plugin=alert,amps,cloud,connections,core,cpu,diskio,docker,folders,fs,gpu,help,ip,irq,load,mem,memswap,network,now,percpu,ports,processcount,processlist,psutilversion,quicklook,raid,sensors,smart,system,uptime,wifi
```
This isn't a total solution, because it's possible a future glances update will introduce a new plugin which is default-on. | non_defect | config to disable all plugins by default or enable an exclusive list thank you for making glances it s awesome and i ve been looking into using it in combination with home assistant to monitor system stats on my network is your feature request related to a problem please describe when deploying glances as a json api only i d like to be able to limit the stats exposed monitoring simple stats like cpu and memory are very useful to me but i would prefer to not expose things like the process list and software versions to the network since they could aid automated vulnerability finders describe the solution you d like it d be really useful to either be able to disable all plugins by default for example support for an argument like disable plugin or disable plugin all would work well for this alternatively it would be useful to be able to specify a list of plugins like plugins cpu mem and have those be the only plugins loaded describe alternatives you ve considered currently the only way i ve found to accomplish this is listing the entire set of available plugins via disable plugin alert amps cloud connections core cpu diskio docker folders fs gpu help ip irq load mem memswap network now percpu ports processcount processlist psutilversion quicklook raid sensors smart system uptime wifi this isn t a total solution because it s possible a future glances update will introduce a new plugin which is default on | 0 |
489,541 | 14,107,902,824 | IssuesEvent | 2020-11-06 16:58:12 | magento/magento2 | https://api.github.com/repos/magento/magento2 | opened | [Issue] Add CartItemOptionsProcessor to Magento_Quote's API | Component: Quote Priority: P3 Severity: S3 | This issue is automatically created based on existing pull request: magento/magento2#30819: Add CartItemOptionsProcessor to Magento_Quote's API
---------
<!---
Thank you for contributing to Magento.
To help us process this pull request we recommend that you add the following information:
- Summary of the pull request,
- Issue(s) related to the changes made,
- Manual testing scenarios
Fields marked with (*) are required. Please don't remove the template.
-->
<!--- Please provide a general summary of the Pull Request in the Title above -->
### Description (*)
This class is used by at least one VBE and is necessary for loading Cart Item Options for quotes that are not "active" (for example, an order in the admin panel)
### Contribution checklist (*)
- [ ] Pull request has a meaningful description of its purpose
- [ ] All commits are accompanied by meaningful commit messages
- [ ] All new or changed code is covered with unit/integration tests (if applicable)
- [ ] All automated tests passed successfully (all builds are green)
| 1.0 | [Issue] Add CartItemOptionsProcessor to Magento_Quote's API - This issue is automatically created based on existing pull request: magento/magento2#30819: Add CartItemOptionsProcessor to Magento_Quote's API
---------
<!---
Thank you for contributing to Magento.
To help us process this pull request we recommend that you add the following information:
- Summary of the pull request,
- Issue(s) related to the changes made,
- Manual testing scenarios
Fields marked with (*) are required. Please don't remove the template.
-->
<!--- Please provide a general summary of the Pull Request in the Title above -->
### Description (*)
This class is used by at least one VBE and is necessary for loading Cart Item Options for quotes that are not "active" (for example, an order in the admin panel)
### Contribution checklist (*)
- [ ] Pull request has a meaningful description of its purpose
- [ ] All commits are accompanied by meaningful commit messages
- [ ] All new or changed code is covered with unit/integration tests (if applicable)
- [ ] All automated tests passed successfully (all builds are green)
| non_defect | add cartitemoptionsprocessor to magento quote s api this issue is automatically created based on existing pull request magento add cartitemoptionsprocessor to magento quote s api thank you for contributing to magento to help us process this pull request we recommend that you add the following information summary of the pull request issue s related to the changes made manual testing scenarios fields marked with are required please don t remove the template description this class is used by at least one vbe and is necessary for loading cart item options for quotes that are not active for example an order in the admin panel contribution checklist pull request has a meaningful description of its purpose all commits are accompanied by meaningful commit messages all new or changed code is covered with unit integration tests if applicable all automated tests passed successfully all builds are green | 0 |
46,522 | 13,055,926,536 | IssuesEvent | 2020-07-30 03:08:31 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | [spline-reco] Segmentation fault in spline reco (Trac #1361) | Incomplete Migration Migrated from Trac combo reconstruction defect | Migrated from https://code.icecube.wisc.edu/ticket/1361
```json
{
"status": "closed",
"changetime": "2015-09-24T09:38:45",
"description": "Segmentation fault shows up when I attempt to run spline-reco module, in the way that is shown in the attachment. I checked trunk and with the last release of icerec, and the same result.",
"reporter": "gmaggi",
"cc": "",
"resolution": "fixed",
"_ts": "1443087525552045",
"component": "combo reconstruction",
"summary": "[spline-reco] Segmentation fault in spline reco",
"priority": "critical",
"keywords": "",
"time": "2015-09-22T06:52:40",
"milestone": "",
"owner": "mvoge",
"type": "defect"
}
```
| 1.0 | [spline-reco] Segmentation fault in spline reco (Trac #1361) - Migrated from https://code.icecube.wisc.edu/ticket/1361
```json
{
"status": "closed",
"changetime": "2015-09-24T09:38:45",
"description": "Segmentation fault shows up when I attempt to run spline-reco module, in the way that is shown in the attachment. I checked trunk and with the last release of icerec, and the same result.",
"reporter": "gmaggi",
"cc": "",
"resolution": "fixed",
"_ts": "1443087525552045",
"component": "combo reconstruction",
"summary": "[spline-reco] Segmentation fault in spline reco",
"priority": "critical",
"keywords": "",
"time": "2015-09-22T06:52:40",
"milestone": "",
"owner": "mvoge",
"type": "defect"
}
```
| defect | segmentation fault in spline reco trac migrated from json status closed changetime description segmentation fault shows up when i attempt to run spline reco module in the way that is shown in the attachment i checked trunk and with the last release of icerec and the same result reporter gmaggi cc resolution fixed ts component combo reconstruction summary segmentation fault in spline reco priority critical keywords time milestone owner mvoge type defect | 1 |
39,406 | 9,437,274,417 | IssuesEvent | 2019-04-13 14:04:08 | UIOWA5830SP19/SPP200 | https://api.github.com/repos/UIOWA5830SP19/SPP200 | closed | User should not like and dislike the same review | T1 - Defect T33 - DB Enhancement | Also, the user should not like and/or dislike the same post multiple times. For resolving this, we would create a new elasticsearch index which would store the review_id, user_id, like/dislike status and last_updated datetime. | 1.0 | User should not like and dislike the same review - Also, the user should not like and/or dislike the same post multiple times. For resolving this, we would create a new elasticsearch index which would store the review_id, user_id, like/dislike status and last_updated datetime. | defect | user should not like and dislike the same review also the user should not like and or dislike the same post multiple times for resolving this we would create a new elasticsearch index which would store the review id user id like dislike status and last updated datetime | 1 |
48,902 | 13,184,770,673 | IssuesEvent | 2020-08-12 20:03:46 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | FIXME's should be fixed. Not warned. (Trac #409) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/409
, reported by nega and owned by karg_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "FIXME's that are warned get ignored and not fixed.\n\n{{{\n[maru:~/i3/icerec/src] svn blame ./tpx/private/tpx/I3IceTopSanityChecks.cxx |grep FIXME \u9f9f trunk-88227\n 78401 kislat #warning FIXME: This is kind of fishy. Subsequent modules should not have to do this check again.\n}}}",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "FIXME's should be fixed. Not warned.",
"priority": "blocker",
"keywords": "tpx",
"time": "2012-05-30T19:28:34",
"milestone": "",
"owner": "karg",
"type": "defect"
}
```
</p>
</details>
| 1.0 | FIXME's should be fixed. Not warned. (Trac #409) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/409
, reported by nega and owned by karg_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "FIXME's that are warned get ignored and not fixed.\n\n{{{\n[maru:~/i3/icerec/src] svn blame ./tpx/private/tpx/I3IceTopSanityChecks.cxx |grep FIXME \u9f9f trunk-88227\n 78401 kislat #warning FIXME: This is kind of fishy. Subsequent modules should not have to do this check again.\n}}}",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "FIXME's should be fixed. Not warned.",
"priority": "blocker",
"keywords": "tpx",
"time": "2012-05-30T19:28:34",
"milestone": "",
"owner": "karg",
"type": "defect"
}
```
</p>
</details>
| defect | fixme s should be fixed not warned trac migrated from reported by nega and owned by karg json status closed changetime description fixme s that are warned get ignored and not fixed n n n svn blame tpx private tpx cxx grep fixme trunk n kislat warning fixme this is kind of fishy subsequent modules should not have to do this check again n reporter nega cc resolution fixed ts component combo reconstruction summary fixme s should be fixed not warned priority blocker keywords tpx time milestone owner karg type defect | 1 |
5,095 | 2,610,180,952 | IssuesEvent | 2015-02-26 18:57:47 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 有关如何治疗脸上色斑 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
寂静的夜,静的似酒令人微醺,都说:夜晚可以使许多人变��
�,诗人,智者,哲学家。我喜欢夜,但变不了诗人,哲学家�
��可思维总跟哲学有关。那一刻,坐在夜色的深处,寂静的怀
里,一种简单的形式,看书写字或冥想,我便成了自己,一��
�真正的自己。那一刻,仿佛有无数双天使的羽翼温柔地呵护�
��我。我可以自由地穿梭于古今,可以任爱恨悲欢汹涌而来;
可以只倾情于一片落叶,一只蚂蚁。可以拥有一个独立的精��
�世界。此刻的夜,不浅薄,此刻的夜,很深沉。此刻,宇宙�
��像一位穿着黑袍神秘的父亲,而我是他多梦的孩子。梦里的
我皮肤光滑,没有烦人的色斑,肌肤像婴儿一样!如何治疗��
�上色斑,
《客户案例》
本人刚踏入社会不久,年龄虽然不大,但是这张脸倒抢��
�了一步“成熟”。大学那时候皮肤到是挺好的,随着岁月的�
��逝,皮肤开始慢慢变黄,光泽度也在下降。而且因为天天都
要面对电脑,脸上慢慢地出现了斑点,越来越多,越来越深��
�大家都说我老了。我很着急,开始美白祛斑产霜、面膜啊什�
��的,但是根本没有效果,还让我脸上起了点疙瘩,过敏变红
,吓得我再也不敢用了。于是我又求助于食物美白祛斑,使��
�喝很多方子,效果比乌龟爬还慢,无奈放弃。</br>
有一天网友给我推荐去黛芙薇尔精华液官网看看,我就��
�开了黛芙薇尔精华液的官网,上面还有在线专家,于是我就�
��专家为什么脸上会有这么多斑,专家告诉我,黛芙薇尔精华
液可清除自由基,击退黑色素,深入色素形成的细胞内发挥��
�接净化黑色素的美白功效,通过各种珍贵活化成分的综合作�
��唤醒人体内自洁酶活性,平衡代谢,解除色素代谢障碍,消
散和色素淤积,增进黑色素自我代谢,听了专家的话我毫不��
�豫的订购了2个周期的黛芙薇尔精华液。</br>
收到产品后,我就开始使用,真是没有想到,一周期用��
�后,斑还真的淡了好多,这莫大的的变化让我打起了精神,�
��更有信心的使用剩下的周期了。用完之后,斑消失了,更让
我欣喜的是皮肤了白嫩了不少,这是没想到以前用了这么多��
�品却没有黛芙薇尔精华液效果来的快,如今的我像是回到了2
0岁,皮肤光滑细嫩,很幸运选择了黛芙薇尔精华液!
阅读了如何治疗脸上色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
如何治疗脸上色斑,同时为您分享祛斑小方法
1、葛根羹
原料:葛粉9克,葡萄干9粒。
做法:将葛粉、葡萄干放入碗中,加少量饮用净水调匀,再��
�沸水冲泡。边冲边搅拌成糊状即可
2、葛根小排汤
原料:葛根100克,山药50克,猪小排250克,食盐2克。
做法:将小排洗净、过水,再放入煮沸的汤水中,加葛根、��
�药同煮,先用旺火再改用文火煲1小时,加入食盐调味即成
以上就是简单介绍了如何美容祛斑,这只能淡化色斑,要想��
�底祛斑还是中药内调好。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 2:25 | 1.0 | 有关如何治疗脸上色斑 - ```
《摘要》
寂静的夜,静的似酒令人微醺,都说:夜晚可以使许多人变��
�,诗人,智者,哲学家。我喜欢夜,但变不了诗人,哲学家�
��可思维总跟哲学有关。那一刻,坐在夜色的深处,寂静的怀
里,一种简单的形式,看书写字或冥想,我便成了自己,一��
�真正的自己。那一刻,仿佛有无数双天使的羽翼温柔地呵护�
��我。我可以自由地穿梭于古今,可以任爱恨悲欢汹涌而来;
可以只倾情于一片落叶,一只蚂蚁。可以拥有一个独立的精��
�世界。此刻的夜,不浅薄,此刻的夜,很深沉。此刻,宇宙�
��像一位穿着黑袍神秘的父亲,而我是他多梦的孩子。梦里的
我皮肤光滑,没有烦人的色斑,肌肤像婴儿一样!如何治疗��
�上色斑,
《客户案例》
本人刚踏入社会不久,年龄虽然不大,但是这张脸倒抢��
�了一步“成熟”。大学那时候皮肤到是挺好的,随着岁月的�
��逝,皮肤开始慢慢变黄,光泽度也在下降。而且因为天天都
要面对电脑,脸上慢慢地出现了斑点,越来越多,越来越深��
�大家都说我老了。我很着急,开始美白祛斑产霜、面膜啊什�
��的,但是根本没有效果,还让我脸上起了点疙瘩,过敏变红
,吓得我再也不敢用了。于是我又求助于食物美白祛斑,使��
�喝很多方子,效果比乌龟爬还慢,无奈放弃。</br>
有一天网友给我推荐去黛芙薇尔精华液官网看看,我就��
�开了黛芙薇尔精华液的官网,上面还有在线专家,于是我就�
��专家为什么脸上会有这么多斑,专家告诉我,黛芙薇尔精华
液可清除自由基,击退黑色素,深入色素形成的细胞内发挥��
�接净化黑色素的美白功效,通过各种珍贵活化成分的综合作�
��唤醒人体内自洁酶活性,平衡代谢,解除色素代谢障碍,消
散和色素淤积,增进黑色素自我代谢,听了专家的话我毫不��
�豫的订购了2个周期的黛芙薇尔精华液。</br>
收到产品后,我就开始使用,真是没有想到,一周期用��
�后,斑还真的淡了好多,这莫大的的变化让我打起了精神,�
��更有信心的使用剩下的周期了。用完之后,斑消失了,更让
我欣喜的是皮肤了白嫩了不少,这是没想到以前用了这么多��
�品却没有黛芙薇尔精华液效果来的快,如今的我像是回到了2
0岁,皮肤光滑细嫩,很幸运选择了黛芙薇尔精华液!
阅读了如何治疗脸上色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
如何治疗脸上色斑,同时为您分享祛斑小方法
1、葛根羹
原料:葛粉9克,葡萄干9粒。
做法:将葛粉、葡萄干放入碗中,加少量饮用净水调匀,再��
�沸水冲泡。边冲边搅拌成糊状即可
2、葛根小排汤
原料:葛根100克,山药50克,猪小排250克,食盐2克。
做法:将小排洗净、过水,再放入煮沸的汤水中,加葛根、��
�药同煮,先用旺火再改用文火煲1小时,加入食盐调味即成
以上就是简单介绍了如何美容祛斑,这只能淡化色斑,要想��
�底祛斑还是中药内调好。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 2:25 | defect | 有关如何治疗脸上色斑 《摘要》 寂静的夜,静的似酒令人微醺,都说:夜晚可以使许多人变�� �,诗人,智者,哲学家。我喜欢夜,但变不了诗人,哲学家� ��可思维总跟哲学有关。那一刻,坐在夜色的深处,寂静的怀 里,一种简单的形式,看书写字或冥想,我便成了自己,一�� �真正的自己。那一刻,仿佛有无数双天使的羽翼温柔地呵护� ��我。我可以自由地穿梭于古今,可以任爱恨悲欢汹涌而来; 可以只倾情于一片落叶,一只蚂蚁。可以拥有一个独立的精�� �世界。此刻的夜,不浅薄,此刻的夜,很深沉。此刻,宇宙� ��像一位穿着黑袍神秘的父亲,而我是他多梦的孩子。梦里的 我皮肤光滑,没有烦人的色斑,肌肤像婴儿一样!如何治疗�� �上色斑, 《客户案例》 本人刚踏入社会不久,年龄虽然不大,但是这张脸倒抢�� �了一步“成熟”。大学那时候皮肤到是挺好的,随着岁月的� ��逝,皮肤开始慢慢变黄,光泽度也在下降。而且因为天天都 要面对电脑,脸上慢慢地出现了斑点,越来越多,越来越深�� �大家都说我老了。我很着急,开始美白祛斑产霜、面膜啊什� ��的,但是根本没有效果,还让我脸上起了点疙瘩,过敏变红 ,吓得我再也不敢用了。于是我又求助于食物美白祛斑,使�� �喝很多方子,效果比乌龟爬还慢,无奈放弃。 有一天网友给我推荐去黛芙薇尔精华液官网看看,我就�� �开了黛芙薇尔精华液的官网,上面还有在线专家,于是我就� ��专家为什么脸上会有这么多斑,专家告诉我,黛芙薇尔精华 液可清除自由基,击退黑色素,深入色素形成的细胞内发挥�� �接净化黑色素的美白功效,通过各种珍贵活化成分的综合作� ��唤醒人体内自洁酶活性,平衡代谢,解除色素代谢障碍,消 散和色素淤积,增进黑色素自我代谢,听了专家的话我毫不�� � 。 收到产品后,我就开始使用,真是没有想到,一周期用�� �后,斑还真的淡了好多,这莫大的的变化让我打起了精神,� ��更有信心的使用剩下的周期了。用完之后,斑消失了,更让 我欣喜的是皮肤了白嫩了不少,这是没想到以前用了这么多�� �品却没有黛芙薇尔精华液效果来的快, ,皮肤光滑细嫩,很幸运选择了黛芙薇尔精华液 阅读了如何治疗脸上色斑,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 如何治疗脸上色斑,同时为您分享祛斑小方法 、葛根羹 原料: , 。 做法:将葛粉、葡萄干放入碗中,加少量饮用净水调匀,再�� �沸水冲泡。边冲边搅拌成糊状即可 、葛根小排汤 原料: , , , 。 做法:将小排洗净、过水,再放入煮沸的汤水中,加葛根、�� �药同煮, ,加入食盐调味即成 以上就是简单介绍了如何美容祛斑,这只能淡化色斑,要想�� �底祛斑还是中药内调好。 original issue reported on code google com by additive gmail com on jul at | 1 |
54,943 | 14,077,350,303 | IssuesEvent | 2020-11-04 11:53:37 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | sendfile regression detected in ZFS 2.0 (tip) with 5.10-rc2 (and not 5.9) | Status: Triage Needed Type: Defect | <!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | Hirsute (development)
Linux Kernel | 5.10-rc2
Architecture | x86-64
ZFS Version | ZFS 2.0 (tip, on commit 71ae6a9d232de59534e4337efb0dc85c7d6906bd)
SPL Version | as above
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
Running xfs test 249 I'm seeing a regression just on the 5.10-rc2 kernel (and not on the 5.9 stable kernel) with ZFS. The failure is in sendfile returning -EINVAL with the 5.10-rc2 kernel where as in 5.9 it works fine.
Just a side note, I back-ported the ZFS 5.10/5.9 compat changes to an earlier 0.8.4 ZFS and I see the same issue, so I'm not sure if this a 5.10-rc2 issue or a ZFS compat issue or a combo of both.
### Describe how to reproduce the problem
Reproducer:
```
#!/bin/bash
VDEV_PATH=${PWD}
POOL=testpool
TESTDIR=test
vdev0=${VDEV_PATH}/block-dev-0
truncate -s 1G ${vdev0}
zpool destroy $POOL &> /dev/null || true
zpool create $POOL $vdev0 -f
zfs create $POOL/$TESTDIR
#zfs set acltype=posixacl $POOL/$TESTDIR
#zfs set compression=on $POOL/$TESTDIR
xfs_io -f -c "pwrite -S 0xa5a55a5a 0 32768k" -c fsync /$POOL/$TESTDIR/249.src
strace -f xfs_io -f -c "sendfile -i /$POOL/$TESTDIR/249.src 0 32768k" -c fsync /$POOL/$TESTDIR/249.dst
zpool destroy $POOL
rm -rf $vdev0
```
With ZFS 2.0 + 5.10-rc2 we see the sendfile return an unexpected -EINVAL as follows:
```
wrote 33554432/33554432 bytes at offset 0
32 MiB, 8192 ops; 0.1738 sec (184.106 MiB/sec and 47131.1122 ops/sec)
execve("/usr/sbin/xfs_io", ["xfs_io", "-f", "-c", "sendfile -i /testpool/test/249.s"..., "-c", "fsync", "/testpool/test/249.dst"], 0x7ffea4835c68 /* 16 vars */) = 0
brk(NULL) = 0x560b3693c000
arch_prctl(0x3001 /* ARCH_??? */, 0x7ffc701a7880) = -1 EINVAL (Invalid argument)
access("/etc/ld.so.preload", R_OK) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0644, st_size=43079, ...}) = 0
mmap(NULL, 43079, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7f9549b1d000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libpthread.so.0", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0000\201\0\0\0\0\0\0"..., 832) = 832
pread64(3, "\4\0\0\0\24\0\0\0\3\0\0\0GNU\0{\324M\31\306%'\265\343\356k%\214\217\2035"..., 68, 824) = 68
fstat(3, {st_mode=S_IFREG|0755, st_size=151232, ...}) = 0
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f9549b1b000
mmap(NULL, 136272, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549af9000
mmap(0x7f9549b00000, 65536, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x7000) = 0x7f9549b00000
mmap(0x7f9549b10000, 20480, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x17000) = 0x7f9549b10000
mmap(0x7f9549b15000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1b000) = 0x7f9549b15000
mmap(0x7f9549b17000, 13392, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549b17000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libreadline.so.5", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\0\34\1\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=253456, ...}) = 0
mmap(NULL, 260536, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549ab9000
mmap(0x7f9549aca000, 122880, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x11000) = 0x7f9549aca000
mmap(0x7f9549ae8000, 32768, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x2f000) = 0x7f9549ae8000
mmap(0x7f9549af0000, 32768, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x36000) = 0x7f9549af0000
mmap(0x7f9549af8000, 2488, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549af8000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libdevmapper.so.1.02.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0@\255\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=431472, ...}) = 0
mmap(NULL, 437648, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549a4e000
mmap(0x7f9549a58000, 294912, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xa000) = 0x7f9549a58000
mmap(0x7f9549aa0000, 81920, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x52000) = 0x7f9549aa0000
mmap(0x7f9549ab4000, 16384, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x65000) = 0x7f9549ab4000
mmap(0x7f9549ab8000, 3472, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549ab8000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libc.so.6", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\3\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\360\215\2\0\0\0\0\0"..., 832) = 832
pread64(3, "\6\0\0\0\4\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0"..., 784, 64) = 784
pread64(3, "\4\0\0\0\20\0\0\0\5\0\0\0GNU\0\2\0\0\300\4\0\0\0\3\0\0\0\0\0\0\0", 32, 848) = 32
pread64(3, "\4\0\0\0\24\0\0\0\3\0\0\0GNU\0~\303\347M\250B\312<j\233\242\v!0<\341"..., 68, 880) = 68
fstat(3, {st_mode=S_IFREG|0755, st_size=1995896, ...}) = 0
pread64(3, "\6\0\0\0\4\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0"..., 784, 64) = 784
mmap(NULL, 2004064, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549864000
mprotect(0x7f954988a000, 1810432, PROT_NONE) = 0
mmap(0x7f954988a000, 1495040, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x26000) = 0x7f954988a000
mmap(0x7f95499f7000, 311296, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x193000) = 0x7f95499f7000
mmap(0x7f9549a44000, 24576, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1df000) = 0x7f9549a44000
mmap(0x7f9549a4a000, 13408, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549a4a000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libtinfo.so.6", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\240\346\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=192032, ...}) = 0
mmap(NULL, 194944, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549834000
mmap(0x7f9549842000, 61440, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xe000) = 0x7f9549842000
mmap(0x7f9549851000, 57344, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1d000) = 0x7f9549851000
mmap(0x7f954985f000, 20480, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x2a000) = 0x7f954985f000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libselinux.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0 p\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=163256, ...}) = 0
mmap(NULL, 174568, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549809000
mmap(0x7f954980f000, 102400, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x6000) = 0x7f954980f000
mmap(0x7f9549828000, 32768, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1f000) = 0x7f9549828000
mmap(0x7f9549830000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x26000) = 0x7f9549830000
mmap(0x7f9549832000, 6632, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549832000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libudev.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0000N\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=158056, ...}) = 0
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f9549807000
mmap(NULL, 162080, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f95497df000
mmap(0x7f95497e3000, 102400, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x4000) = 0x7f95497e3000
mmap(0x7f95497fc000, 36864, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1d000) = 0x7f95497fc000
mmap(0x7f9549805000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x25000) = 0x7f9549805000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libm.so.6", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\3\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\300\363\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=1369352, ...}) = 0
mmap(NULL, 1368336, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549690000
mmap(0x7f954969f000, 684032, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xf000) = 0x7f954969f000
mmap(0x7f9549746000, 618496, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xb6000) = 0x7f9549746000
mmap(0x7f95497dd000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x14c000) = 0x7f95497dd000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libpcre2-8.so.0", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\340\"\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=584392, ...}) = 0
mmap(NULL, 586536, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549600000
mmap(0x7f9549602000, 409600, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x2000) = 0x7f9549602000
mmap(0x7f9549666000, 163840, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x66000) = 0x7f9549666000
mmap(0x7f954968e000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x8d000) = 0x7f954968e000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libdl.so.2", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0 \22\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=18816, ...}) = 0
mmap(NULL, 20752, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f95495fa000
mmap(0x7f95495fb000, 8192, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1000) = 0x7f95495fb000
mmap(0x7f95495fd000, 4096, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x3000) = 0x7f95495fd000
mmap(0x7f95495fe000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x3000) = 0x7f95495fe000
close(3) = 0
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f95495f8000
mmap(NULL, 12288, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f95495f5000
arch_prctl(ARCH_SET_FS, 0x7f95495f5800) = 0
mprotect(0x7f9549a44000, 12288, PROT_READ) = 0
mprotect(0x7f95495fe000, 4096, PROT_READ) = 0
mprotect(0x7f9549b15000, 4096, PROT_READ) = 0
mprotect(0x7f954968e000, 4096, PROT_READ) = 0
mprotect(0x7f95497dd000, 4096, PROT_READ) = 0
mprotect(0x7f9549805000, 4096, PROT_READ) = 0
mprotect(0x7f9549830000, 4096, PROT_READ) = 0
mprotect(0x7f954985f000, 16384, PROT_READ) = 0
mprotect(0x7f9549ab4000, 4096, PROT_READ) = 0
mprotect(0x7f9549af0000, 8192, PROT_READ) = 0
mprotect(0x560b354c6000, 4096, PROT_READ) = 0
mprotect(0x7f9549b56000, 4096, PROT_READ) = 0
munmap(0x7f9549b1d000, 43079) = 0
set_tid_address(0x7f95495f5ad0) = 3265
set_robust_list(0x7f95495f5ae0, 24) = 0
rt_sigaction(SIGRTMIN, {sa_handler=0x7f9549b00b90, sa_mask=[], sa_flags=SA_RESTORER|SA_SIGINFO, sa_restorer=0x7f9549b0dbb0}, NULL, 8) = 0
rt_sigaction(SIGRT_1, {sa_handler=0x7f9549b00c30, sa_mask=[], sa_flags=SA_RESTORER|SA_RESTART|SA_SIGINFO, sa_restorer=0x7f9549b0dbb0}, NULL, 8) = 0
rt_sigprocmask(SIG_UNBLOCK, [RTMIN RT_1], NULL, 8) = 0
prlimit64(0, RLIMIT_STACK, NULL, {rlim_cur=8192*1024, rlim_max=RLIM64_INFINITY}) = 0
statfs("/sys/fs/selinux", 0x7ffc701a77d0) = -1 ENOENT (No such file or directory)
statfs("/selinux", 0x7ffc701a77d0) = -1 ENOENT (No such file or directory)
brk(NULL) = 0x560b3693c000
brk(0x560b3695d000) = 0x560b3695d000
openat(AT_FDCWD, "/proc/filesystems", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0444, st_size=0, ...}) = 0
read(3, "nodev\tsysfs\nnodev\ttmpfs\nnodev\tbd"..., 1024) = 420
read(3, "", 1024) = 0
close(3) = 0
access("/etc/selinux/config", F_OK) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/lib/locale/locale-archive", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0644, st_size=3041504, ...}) = 0
mmap(NULL, 3041504, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7f954930e000
close(3) = 0
access("/proc/self/mounts", R_OK) = 0
openat(AT_FDCWD, "/proc/self/mounts", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0444, st_size=0, ...}) = 0
read(3, "sysfs /sys sysfs rw,nosuid,nodev"..., 1024) = 1024
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/sysfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/proc", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/proc", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/udev", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/pts", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/devpts", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/vda2", {st_mode=S_IFBLK|0660, st_rdev=makedev(0xfc, 0x2), ...}) = 0
stat("/", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
statfs("/", {f_type=EXT2_SUPER_MAGIC, f_bsize=4096, f_blocks=7578609, f_bfree=3819671, f_bavail=3428939, f_files=1933312, f_ffree=1686992, f_fsid={val=[3617647594, 2305891747]}, f_namelen=255, f_frsize=4096, f_flags=ST_VALID|ST_RELATIME}) = 0
stat("/", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/security", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/securityfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/shm", {st_mode=S_IFDIR|S_ISVTX|0777, st_size=40, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/lock", {st_mode=S_IFDIR|S_ISVTX|0777, st_size=80, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/unified", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup2", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/systemd", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/pstore", {st_mode=S_IFDIR|0750, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/pstore", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/firmware", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/firmware/efi", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/firmware/efi/efivars", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/efivarfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
read(3, "xec,relatime,mode=700 0 0\ncgroup"..., 1024) = 1024
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/bpf", {st_mode=S_IFDIR|S_ISVTX|0700, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/none", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/cpu,cpuacct", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/net_cls,net_prio", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/rdma", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/devices", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/perf_event", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/freezer", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/memory", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/pids", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/cpuset", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/hugetlb", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/blkio", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
read(3, "o=5,direct,pipe_ino=18625 0 0\nhu"..., 1024) = 1024
lstat("/proc", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/sys/fs", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/sys/fs/binfmt_misc", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/systemd-1", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/hugepages", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/hugetlbfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/mqueue", {st_mode=S_IFDIR|S_ISVTX|0777, st_size=40, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/mqueue", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/debug", {st_mode=S_IFDIR|0700, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/debugfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/tracing", {st_mode=S_IFDIR|0700, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tracefs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/rpc_pipefs", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/sunrpc", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/proc", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/fs", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/fs/nfsd", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/nfsd", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/fuse", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/fuse/connections", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/fusectl", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/config", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/configfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/boot", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/boot/efi", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/vda1", {st_mode=S_IFBLK|0660, st_rdev=makedev(0xfc, 0x1), ...}) = 0
stat("/boot/efi", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
statfs("/boot/efi", {f_type=MSDOS_SUPER_MAGIC, f_bsize=4096, f_blocks=130812, f_bfree=128816, f_bavail=128816, f_files=0, f_ffree=0, f_fsid={val=[64513, 0]}, f_namelen=1530, f_frsize=4096, f_flags=ST_VALID|ST_RELATIME}) = 0
stat("/boot/efi", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18/1932", {st_mode=S_IFDIR|0755, st_size=348, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop1", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x1), ...}) = 0
stat("/snap/core18/1932", {st_mode=S_IFDIR|0755, st_size=348, ...}) = 0
statfs("/snap/core18/1932", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=443, f_bfree=0, f_bavail=0, f_files=10779, f_ffree=0, f_fsid={val=[1793, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/core18/1932", {st_mode=S_IFDIR|0755, st_size=348, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18/1885", {st_mode=S_IFDIR|0755, st_size=334, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop0", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0), ...}) = 0
stat("/snap/core18/1885", {st_mode=S_IFDIR|0755, st_size=334, ...}) = 0
statfs("/snap/core18/1885", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=443, f_bfree=0, f_bavail=0, f_files=10775, f_ffree=0, f_fsid={val=[1792, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/core18/1885", {st_mode=S_IFDIR|0755, st_size=334, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd/17936", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop2", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x2), ...}) = 0
stat("/snap/lxd/17936", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
statfs("/snap/lxd/17936", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=554, f_bfree=0, f_bavail=0, f_files=1559, f_ffree=0, f_fsid={val=[1794, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/lxd/17936", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd/18137", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop3", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x3), ...}) = 0
stat("/snap/lxd/18137", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
statfs("/snap/lxd/18137", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=554, f_bfree=0, f_bavail=0, f_files=1559, f_ffree=0, f_fsid={val=[1795, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/lxd/18137", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/snapd", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/snapd/9721", {st_mode=S_IFDIR|0755, st_size=119, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop4", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x4), ...}) = 0
stat("/snap/snapd/9721", {st_mode=S_IFDIR|0755, st_size=119, ...}) = 0
statfs("/snap/snapd/9721", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=248, f_bfree=0, f_bavail=0, f_files=472, f_ffree=0, f_fsid={val=[1796, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/snapd/9721", {st_mode=S_IFDIR|0755, st_size=119, ...}) = 0
read(3, "=401880k,mode=755 0 0\nnsfs /run/"..., 1024) = 268
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/snapd", {st_mode=S_IFDIR|0755, st_size=80, ...}) = 0
lstat("/run/snapd/ns", {st_mode=S_IFDIR|0755, st_size=100, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/snapd", {st_mode=S_IFDIR|0755, st_size=80, ...}) = 0
lstat("/run/snapd/ns", {st_mode=S_IFDIR|0755, st_size=100, ...}) = 0
lstat("/run/snapd/ns/lxd.mnt", {st_mode=S_IFREG|0444, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/nsfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/user", {st_mode=S_IFDIR|0755, st_size=60, ...}) = 0
lstat("/run/user/1000", {st_mode=S_IFDIR|0700, st_size=140, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/testpool", {st_mode=S_IFDIR|0755, st_size=3, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/testpool", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/testpool", {st_mode=S_IFDIR|0755, st_size=3, ...}) = 0
lstat("/testpool/test", {st_mode=S_IFDIR|0755, st_size=3, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/testpool", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
read(3, "", 1024) = 0
close(3) = 0
openat(AT_FDCWD, "/etc/projects", O_RDONLY) = -1 ENOENT (No such file or directory)
stat("/testpool/test/249.dst", 0x7ffc701a73f0) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/testpool/test/249.dst", O_RDWR|O_CREAT, 0600) = 3
fstatfs(3, {f_type=ZFS_SUPER_MAGIC, f_bsize=131072, f_blocks=6655, f_bfree=6398, f_bavail=6398, f_files=1638096, f_ffree=1638089, f_fsid={val=[1239261780, 15049724]}, f_namelen=255, f_frsize=131072, f_flags=ST_VALID}) = 0
fstat(3, {st_mode=S_IFREG|0600, st_size=0, ...}) = 0
fstatfs(3, {f_type=ZFS_SUPER_MAGIC, f_bsize=131072, f_blocks=6655, f_bfree=6398, f_bavail=6398, f_files=1638096, f_ffree=1638089, f_fsid={val=[1239261780, 15049724]}, f_namelen=255, f_frsize=131072, f_flags=ST_VALID}) = 0
fstat(3, {st_mode=S_IFREG|0600, st_size=0, ...}) = 0
openat(AT_FDCWD, "/usr/share/locale/locale.alias", O_RDONLY|O_CLOEXEC) = 4
fstat(4, {st_mode=S_IFREG|0644, st_size=2996, ...}) = 0
read(4, "# Locale name alias data base.\n#"..., 4096) = 2996
read(4, "", 4096) = 0
close(4) = 0
openat(AT_FDCWD, "/usr/share/locale/en_GB.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
sysinfo({uptime=626, loads=[4256, 1472, 224], totalram=4115226624, freeram=3357290496, sharedram=1187840, bufferram=43425792, totalswap=4116705280, freeswap=4116705280, procs=278, totalhigh=0, freehigh=0, mem_unit=1}) = 0
stat("/testpool/test/249.src", {st_mode=S_IFREG|0600, st_size=33554432, ...}) = 0
openat(AT_FDCWD, "/testpool/test/249.src", O_RDONLY) = 4
sendfile(3, 4, [0], 33554432) = -1 EINVAL (Invalid argument)
dup(2) = 5
fcntl(5, F_GETFL) = 0x2 (flags O_RDWR)
openat(AT_FDCWD, "/usr/share/locale/en_GB.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
fstat(5, {st_mode=S_IFCHR|0620, st_rdev=makedev(0x88, 0), ...}) = 0
write(5, "sendfile: Invalid argument\n", 27sendfile: Invalid argument
) = 27
close(5) = 0
close(4) = 0
fsync(3) = 0
exit_group(1) = ?
+++ exited with 1 +++
```
This works fine on 5.9 with ZFS 2.0. | 1.0 | sendfile regression detected in ZFS 2.0 (tip) with 5.10-rc2 (and not 5.9) - <!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | Hirsute (development)
Linux Kernel | 5.10-rc2
Architecture | x86-64
ZFS Version | ZFS 2.0 (tip, on commit 71ae6a9d232de59534e4337efb0dc85c7d6906bd)
SPL Version | as above
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
Running xfs test 249 I'm seeing a regression just on the 5.10-rc2 kernel (and not on the 5.9 stable kernel) with ZFS. The failure is in sendfile returning -EINVAL with the 5.10-rc2 kernel where as in 5.9 it works fine.
Just a side note, I back-ported the ZFS 5.10/5.9 compat changes to an earlier 0.8.4 ZFS and I see the same issue, so I'm not sure if this a 5.10-rc2 issue or a ZFS compat issue or a combo of both.
### Describe how to reproduce the problem
Reproducer:
```
#!/bin/bash
VDEV_PATH=${PWD}
POOL=testpool
TESTDIR=test
vdev0=${VDEV_PATH}/block-dev-0
truncate -s 1G ${vdev0}
zpool destroy $POOL &> /dev/null || true
zpool create $POOL $vdev0 -f
zfs create $POOL/$TESTDIR
#zfs set acltype=posixacl $POOL/$TESTDIR
#zfs set compression=on $POOL/$TESTDIR
xfs_io -f -c "pwrite -S 0xa5a55a5a 0 32768k" -c fsync /$POOL/$TESTDIR/249.src
strace -f xfs_io -f -c "sendfile -i /$POOL/$TESTDIR/249.src 0 32768k" -c fsync /$POOL/$TESTDIR/249.dst
zpool destroy $POOL
rm -rf $vdev0
```
With ZFS 2.0 + 5.10-rc2 we see the sendfile return an unexpected -EINVAL as follows:
```
wrote 33554432/33554432 bytes at offset 0
32 MiB, 8192 ops; 0.1738 sec (184.106 MiB/sec and 47131.1122 ops/sec)
execve("/usr/sbin/xfs_io", ["xfs_io", "-f", "-c", "sendfile -i /testpool/test/249.s"..., "-c", "fsync", "/testpool/test/249.dst"], 0x7ffea4835c68 /* 16 vars */) = 0
brk(NULL) = 0x560b3693c000
arch_prctl(0x3001 /* ARCH_??? */, 0x7ffc701a7880) = -1 EINVAL (Invalid argument)
access("/etc/ld.so.preload", R_OK) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0644, st_size=43079, ...}) = 0
mmap(NULL, 43079, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7f9549b1d000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libpthread.so.0", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0000\201\0\0\0\0\0\0"..., 832) = 832
pread64(3, "\4\0\0\0\24\0\0\0\3\0\0\0GNU\0{\324M\31\306%'\265\343\356k%\214\217\2035"..., 68, 824) = 68
fstat(3, {st_mode=S_IFREG|0755, st_size=151232, ...}) = 0
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f9549b1b000
mmap(NULL, 136272, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549af9000
mmap(0x7f9549b00000, 65536, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x7000) = 0x7f9549b00000
mmap(0x7f9549b10000, 20480, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x17000) = 0x7f9549b10000
mmap(0x7f9549b15000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1b000) = 0x7f9549b15000
mmap(0x7f9549b17000, 13392, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549b17000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libreadline.so.5", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\0\34\1\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=253456, ...}) = 0
mmap(NULL, 260536, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549ab9000
mmap(0x7f9549aca000, 122880, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x11000) = 0x7f9549aca000
mmap(0x7f9549ae8000, 32768, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x2f000) = 0x7f9549ae8000
mmap(0x7f9549af0000, 32768, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x36000) = 0x7f9549af0000
mmap(0x7f9549af8000, 2488, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549af8000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libdevmapper.so.1.02.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0@\255\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=431472, ...}) = 0
mmap(NULL, 437648, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549a4e000
mmap(0x7f9549a58000, 294912, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xa000) = 0x7f9549a58000
mmap(0x7f9549aa0000, 81920, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x52000) = 0x7f9549aa0000
mmap(0x7f9549ab4000, 16384, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x65000) = 0x7f9549ab4000
mmap(0x7f9549ab8000, 3472, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549ab8000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libc.so.6", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\3\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\360\215\2\0\0\0\0\0"..., 832) = 832
pread64(3, "\6\0\0\0\4\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0"..., 784, 64) = 784
pread64(3, "\4\0\0\0\20\0\0\0\5\0\0\0GNU\0\2\0\0\300\4\0\0\0\3\0\0\0\0\0\0\0", 32, 848) = 32
pread64(3, "\4\0\0\0\24\0\0\0\3\0\0\0GNU\0~\303\347M\250B\312<j\233\242\v!0<\341"..., 68, 880) = 68
fstat(3, {st_mode=S_IFREG|0755, st_size=1995896, ...}) = 0
pread64(3, "\6\0\0\0\4\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0@\0\0\0\0\0\0\0"..., 784, 64) = 784
mmap(NULL, 2004064, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549864000
mprotect(0x7f954988a000, 1810432, PROT_NONE) = 0
mmap(0x7f954988a000, 1495040, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x26000) = 0x7f954988a000
mmap(0x7f95499f7000, 311296, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x193000) = 0x7f95499f7000
mmap(0x7f9549a44000, 24576, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1df000) = 0x7f9549a44000
mmap(0x7f9549a4a000, 13408, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549a4a000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libtinfo.so.6", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\240\346\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=192032, ...}) = 0
mmap(NULL, 194944, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549834000
mmap(0x7f9549842000, 61440, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xe000) = 0x7f9549842000
mmap(0x7f9549851000, 57344, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1d000) = 0x7f9549851000
mmap(0x7f954985f000, 20480, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x2a000) = 0x7f954985f000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libselinux.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0 p\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=163256, ...}) = 0
mmap(NULL, 174568, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549809000
mmap(0x7f954980f000, 102400, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x6000) = 0x7f954980f000
mmap(0x7f9549828000, 32768, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1f000) = 0x7f9549828000
mmap(0x7f9549830000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x26000) = 0x7f9549830000
mmap(0x7f9549832000, 6632, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0) = 0x7f9549832000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libudev.so.1", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0000N\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=158056, ...}) = 0
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f9549807000
mmap(NULL, 162080, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f95497df000
mmap(0x7f95497e3000, 102400, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x4000) = 0x7f95497e3000
mmap(0x7f95497fc000, 36864, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1d000) = 0x7f95497fc000
mmap(0x7f9549805000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x25000) = 0x7f9549805000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libm.so.6", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\3\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\300\363\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=1369352, ...}) = 0
mmap(NULL, 1368336, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549690000
mmap(0x7f954969f000, 684032, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xf000) = 0x7f954969f000
mmap(0x7f9549746000, 618496, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0xb6000) = 0x7f9549746000
mmap(0x7f95497dd000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x14c000) = 0x7f95497dd000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libpcre2-8.so.0", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0\340\"\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=584392, ...}) = 0
mmap(NULL, 586536, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f9549600000
mmap(0x7f9549602000, 409600, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x2000) = 0x7f9549602000
mmap(0x7f9549666000, 163840, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x66000) = 0x7f9549666000
mmap(0x7f954968e000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x8d000) = 0x7f954968e000
close(3) = 0
openat(AT_FDCWD, "/lib/x86_64-linux-gnu/libdl.so.2", O_RDONLY|O_CLOEXEC) = 3
read(3, "\177ELF\2\1\1\0\0\0\0\0\0\0\0\0\3\0>\0\1\0\0\0 \22\0\0\0\0\0\0"..., 832) = 832
fstat(3, {st_mode=S_IFREG|0644, st_size=18816, ...}) = 0
mmap(NULL, 20752, PROT_READ, MAP_PRIVATE|MAP_DENYWRITE, 3, 0) = 0x7f95495fa000
mmap(0x7f95495fb000, 8192, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x1000) = 0x7f95495fb000
mmap(0x7f95495fd000, 4096, PROT_READ, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x3000) = 0x7f95495fd000
mmap(0x7f95495fe000, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_FIXED|MAP_DENYWRITE, 3, 0x3000) = 0x7f95495fe000
close(3) = 0
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f95495f8000
mmap(NULL, 12288, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7f95495f5000
arch_prctl(ARCH_SET_FS, 0x7f95495f5800) = 0
mprotect(0x7f9549a44000, 12288, PROT_READ) = 0
mprotect(0x7f95495fe000, 4096, PROT_READ) = 0
mprotect(0x7f9549b15000, 4096, PROT_READ) = 0
mprotect(0x7f954968e000, 4096, PROT_READ) = 0
mprotect(0x7f95497dd000, 4096, PROT_READ) = 0
mprotect(0x7f9549805000, 4096, PROT_READ) = 0
mprotect(0x7f9549830000, 4096, PROT_READ) = 0
mprotect(0x7f954985f000, 16384, PROT_READ) = 0
mprotect(0x7f9549ab4000, 4096, PROT_READ) = 0
mprotect(0x7f9549af0000, 8192, PROT_READ) = 0
mprotect(0x560b354c6000, 4096, PROT_READ) = 0
mprotect(0x7f9549b56000, 4096, PROT_READ) = 0
munmap(0x7f9549b1d000, 43079) = 0
set_tid_address(0x7f95495f5ad0) = 3265
set_robust_list(0x7f95495f5ae0, 24) = 0
rt_sigaction(SIGRTMIN, {sa_handler=0x7f9549b00b90, sa_mask=[], sa_flags=SA_RESTORER|SA_SIGINFO, sa_restorer=0x7f9549b0dbb0}, NULL, 8) = 0
rt_sigaction(SIGRT_1, {sa_handler=0x7f9549b00c30, sa_mask=[], sa_flags=SA_RESTORER|SA_RESTART|SA_SIGINFO, sa_restorer=0x7f9549b0dbb0}, NULL, 8) = 0
rt_sigprocmask(SIG_UNBLOCK, [RTMIN RT_1], NULL, 8) = 0
prlimit64(0, RLIMIT_STACK, NULL, {rlim_cur=8192*1024, rlim_max=RLIM64_INFINITY}) = 0
statfs("/sys/fs/selinux", 0x7ffc701a77d0) = -1 ENOENT (No such file or directory)
statfs("/selinux", 0x7ffc701a77d0) = -1 ENOENT (No such file or directory)
brk(NULL) = 0x560b3693c000
brk(0x560b3695d000) = 0x560b3695d000
openat(AT_FDCWD, "/proc/filesystems", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0444, st_size=0, ...}) = 0
read(3, "nodev\tsysfs\nnodev\ttmpfs\nnodev\tbd"..., 1024) = 420
read(3, "", 1024) = 0
close(3) = 0
access("/etc/selinux/config", F_OK) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/lib/locale/locale-archive", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0644, st_size=3041504, ...}) = 0
mmap(NULL, 3041504, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7f954930e000
close(3) = 0
access("/proc/self/mounts", R_OK) = 0
openat(AT_FDCWD, "/proc/self/mounts", O_RDONLY|O_CLOEXEC) = 3
fstat(3, {st_mode=S_IFREG|0444, st_size=0, ...}) = 0
read(3, "sysfs /sys sysfs rw,nosuid,nodev"..., 1024) = 1024
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/sysfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/proc", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/proc", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/udev", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/pts", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/devpts", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/vda2", {st_mode=S_IFBLK|0660, st_rdev=makedev(0xfc, 0x2), ...}) = 0
stat("/", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
statfs("/", {f_type=EXT2_SUPER_MAGIC, f_bsize=4096, f_blocks=7578609, f_bfree=3819671, f_bavail=3428939, f_files=1933312, f_ffree=1686992, f_fsid={val=[3617647594, 2305891747]}, f_namelen=255, f_frsize=4096, f_flags=ST_VALID|ST_RELATIME}) = 0
stat("/", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/security", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/securityfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/shm", {st_mode=S_IFDIR|S_ISVTX|0777, st_size=40, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/lock", {st_mode=S_IFDIR|S_ISVTX|0777, st_size=80, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/unified", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup2", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/systemd", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/pstore", {st_mode=S_IFDIR|0750, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/pstore", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/firmware", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/firmware/efi", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/firmware/efi/efivars", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/efivarfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
read(3, "xec,relatime,mode=700 0 0\ncgroup"..., 1024) = 1024
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/bpf", {st_mode=S_IFDIR|S_ISVTX|0700, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/none", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/cpu,cpuacct", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/net_cls,net_prio", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/rdma", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/devices", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/perf_event", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/freezer", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/memory", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/pids", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/cpuset", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/hugetlb", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/cgroup", {st_mode=S_IFDIR|0755, st_size=380, ...}) = 0
lstat("/sys/fs/cgroup/blkio", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/cgroup", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
read(3, "o=5,direct,pipe_ino=18625 0 0\nhu"..., 1024) = 1024
lstat("/proc", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/sys/fs", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/sys/fs/binfmt_misc", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/systemd-1", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/hugepages", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/hugetlbfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/mqueue", {st_mode=S_IFDIR|S_ISVTX|0777, st_size=40, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/mqueue", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/debug", {st_mode=S_IFDIR|0700, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/debugfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/tracing", {st_mode=S_IFDIR|0700, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tracefs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/rpc_pipefs", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/sunrpc", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/proc", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/fs", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/proc/fs/nfsd", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/nfsd", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/fs", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/fuse", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/fs/fuse/connections", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/fusectl", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/sys", {st_mode=S_IFDIR|0555, st_size=0, ...}) = 0
lstat("/sys/kernel", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
lstat("/sys/kernel/config", {st_mode=S_IFDIR|0755, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/configfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/boot", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/boot/efi", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/vda1", {st_mode=S_IFBLK|0660, st_rdev=makedev(0xfc, 0x1), ...}) = 0
stat("/boot/efi", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
statfs("/boot/efi", {f_type=MSDOS_SUPER_MAGIC, f_bsize=4096, f_blocks=130812, f_bfree=128816, f_bavail=128816, f_files=0, f_ffree=0, f_fsid={val=[64513, 0]}, f_namelen=1530, f_frsize=4096, f_flags=ST_VALID|ST_RELATIME}) = 0
stat("/boot/efi", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18/1932", {st_mode=S_IFDIR|0755, st_size=348, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop1", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x1), ...}) = 0
stat("/snap/core18/1932", {st_mode=S_IFDIR|0755, st_size=348, ...}) = 0
statfs("/snap/core18/1932", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=443, f_bfree=0, f_bavail=0, f_files=10779, f_ffree=0, f_fsid={val=[1793, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/core18/1932", {st_mode=S_IFDIR|0755, st_size=348, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/core18/1885", {st_mode=S_IFDIR|0755, st_size=334, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop0", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0), ...}) = 0
stat("/snap/core18/1885", {st_mode=S_IFDIR|0755, st_size=334, ...}) = 0
statfs("/snap/core18/1885", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=443, f_bfree=0, f_bavail=0, f_files=10775, f_ffree=0, f_fsid={val=[1792, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/core18/1885", {st_mode=S_IFDIR|0755, st_size=334, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd/17936", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop2", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x2), ...}) = 0
stat("/snap/lxd/17936", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
statfs("/snap/lxd/17936", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=554, f_bfree=0, f_bavail=0, f_files=1559, f_ffree=0, f_fsid={val=[1794, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/lxd/17936", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/lxd/18137", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop3", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x3), ...}) = 0
stat("/snap/lxd/18137", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
statfs("/snap/lxd/18137", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=554, f_bfree=0, f_bavail=0, f_files=1559, f_ffree=0, f_fsid={val=[1795, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/lxd/18137", {st_mode=S_IFDIR|0755, st_size=246, ...}) = 0
lstat("/snap", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/snapd", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0
lstat("/snap/snapd/9721", {st_mode=S_IFDIR|0755, st_size=119, ...}) = 0
lstat("/dev", {st_mode=S_IFDIR|0755, st_size=4060, ...}) = 0
lstat("/dev/loop4", {st_mode=S_IFBLK|0660, st_rdev=makedev(0x7, 0x4), ...}) = 0
stat("/snap/snapd/9721", {st_mode=S_IFDIR|0755, st_size=119, ...}) = 0
statfs("/snap/snapd/9721", {f_type=SQUASHFS_MAGIC, f_bsize=131072, f_blocks=248, f_bfree=0, f_bavail=0, f_files=472, f_ffree=0, f_fsid={val=[1796, 0]}, f_namelen=256, f_frsize=131072, f_flags=ST_VALID|ST_RDONLY|ST_NODEV|ST_RELATIME}) = 0
stat("/snap/snapd/9721", {st_mode=S_IFDIR|0755, st_size=119, ...}) = 0
read(3, "=401880k,mode=755 0 0\nnsfs /run/"..., 1024) = 268
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/snapd", {st_mode=S_IFDIR|0755, st_size=80, ...}) = 0
lstat("/run/snapd/ns", {st_mode=S_IFDIR|0755, st_size=100, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/snapd", {st_mode=S_IFDIR|0755, st_size=80, ...}) = 0
lstat("/run/snapd/ns", {st_mode=S_IFDIR|0755, st_size=100, ...}) = 0
lstat("/run/snapd/ns/lxd.mnt", {st_mode=S_IFREG|0444, st_size=0, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/nsfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/run", {st_mode=S_IFDIR|0755, st_size=920, ...}) = 0
lstat("/run/user", {st_mode=S_IFDIR|0755, st_size=60, ...}) = 0
lstat("/run/user/1000", {st_mode=S_IFDIR|0700, st_size=140, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/tmpfs", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/testpool", {st_mode=S_IFDIR|0755, st_size=3, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/testpool", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
lstat("/testpool", {st_mode=S_IFDIR|0755, st_size=3, ...}) = 0
lstat("/testpool/test", {st_mode=S_IFDIR|0755, st_size=3, ...}) = 0
getcwd("/home/cking", 4096) = 12
lstat("/home/cking/testpool", 0x7ffc701a4390) = -1 ENOENT (No such file or directory)
read(3, "", 1024) = 0
close(3) = 0
openat(AT_FDCWD, "/etc/projects", O_RDONLY) = -1 ENOENT (No such file or directory)
stat("/testpool/test/249.dst", 0x7ffc701a73f0) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/testpool/test/249.dst", O_RDWR|O_CREAT, 0600) = 3
fstatfs(3, {f_type=ZFS_SUPER_MAGIC, f_bsize=131072, f_blocks=6655, f_bfree=6398, f_bavail=6398, f_files=1638096, f_ffree=1638089, f_fsid={val=[1239261780, 15049724]}, f_namelen=255, f_frsize=131072, f_flags=ST_VALID}) = 0
fstat(3, {st_mode=S_IFREG|0600, st_size=0, ...}) = 0
fstatfs(3, {f_type=ZFS_SUPER_MAGIC, f_bsize=131072, f_blocks=6655, f_bfree=6398, f_bavail=6398, f_files=1638096, f_ffree=1638089, f_fsid={val=[1239261780, 15049724]}, f_namelen=255, f_frsize=131072, f_flags=ST_VALID}) = 0
fstat(3, {st_mode=S_IFREG|0600, st_size=0, ...}) = 0
openat(AT_FDCWD, "/usr/share/locale/locale.alias", O_RDONLY|O_CLOEXEC) = 4
fstat(4, {st_mode=S_IFREG|0644, st_size=2996, ...}) = 0
read(4, "# Locale name alias data base.\n#"..., 4096) = 2996
read(4, "", 4096) = 0
close(4) = 0
openat(AT_FDCWD, "/usr/share/locale/en_GB.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.UTF-8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.utf8/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en/LC_MESSAGES/xfsprogs.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
sysinfo({uptime=626, loads=[4256, 1472, 224], totalram=4115226624, freeram=3357290496, sharedram=1187840, bufferram=43425792, totalswap=4116705280, freeswap=4116705280, procs=278, totalhigh=0, freehigh=0, mem_unit=1}) = 0
stat("/testpool/test/249.src", {st_mode=S_IFREG|0600, st_size=33554432, ...}) = 0
openat(AT_FDCWD, "/testpool/test/249.src", O_RDONLY) = 4
sendfile(3, 4, [0], 33554432) = -1 EINVAL (Invalid argument)
dup(2) = 5
fcntl(5, F_GETFL) = 0x2 (flags O_RDWR)
openat(AT_FDCWD, "/usr/share/locale/en_GB.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en_GB/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale/en/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en_GB/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.UTF-8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en.utf8/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
openat(AT_FDCWD, "/usr/share/locale-langpack/en/LC_MESSAGES/libc.mo", O_RDONLY) = -1 ENOENT (No such file or directory)
fstat(5, {st_mode=S_IFCHR|0620, st_rdev=makedev(0x88, 0), ...}) = 0
write(5, "sendfile: Invalid argument\n", 27sendfile: Invalid argument
) = 27
close(5) = 0
close(4) = 0
fsync(3) = 0
exit_group(1) = ?
+++ exited with 1 +++
```
This works fine on 5.9 with ZFS 2.0. | defect | sendfile regression detected in zfs tip with and not type version name distribution name ubuntu distribution version hirsute development linux kernel architecture zfs version zfs tip on commit spl version as above commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version running xfs test i m seeing a regression just on the kernel and not on the stable kernel with zfs the failure is in sendfile returning einval with the kernel where as in it works fine just a side note i back ported the zfs compat changes to an earlier zfs and i see the same issue so i m not sure if this a issue or a zfs compat issue or a combo of both describe how to reproduce the problem reproducer bin bash vdev path pwd pool testpool testdir test vdev path block dev truncate s zpool destroy pool dev null true zpool create pool f zfs create pool testdir zfs set acltype posixacl pool testdir zfs set compression on pool testdir xfs io f c pwrite s c fsync pool testdir src strace f xfs io f c sendfile i pool testdir src c fsync pool testdir dst zpool destroy pool rm rf with zfs we see the sendfile return an unexpected einval as follows wrote bytes at offset mib ops sec mib sec and ops sec execve usr sbin xfs io vars brk null arch prctl arch einval invalid argument access etc ld so preload r ok enoent no such file or directory openat at fdcwd etc ld so cache o rdonly o cloexec fstat st mode s ifreg st size mmap null prot read map private close openat at fdcwd lib linux gnu libpthread so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read prot write map private map anonymous mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite mmap prot read prot write map private map fixed map anonymous close openat at fdcwd lib linux gnu libreadline so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite mmap prot read prot write map private map fixed map anonymous close openat at fdcwd lib linux gnu libdevmapper so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite mmap prot read prot write map private map fixed map anonymous close openat at fdcwd lib linux gnu libc so o rdonly o cloexec read j v fstat st mode s ifreg st size mmap null prot read map private map denywrite mprotect prot none mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite mmap prot read prot write map private map fixed map anonymous close openat at fdcwd lib linux gnu libtinfo so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite close openat at fdcwd lib linux gnu libselinux so o rdonly o cloexec read p fstat st mode s ifreg st size mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite mmap prot read prot write map private map fixed map anonymous close openat at fdcwd lib linux gnu libudev so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read prot write map private map anonymous mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite close openat at fdcwd lib linux gnu libm so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite close openat at fdcwd lib linux gnu so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite close openat at fdcwd lib linux gnu libdl so o rdonly o cloexec read fstat st mode s ifreg st size mmap null prot read map private map denywrite mmap prot read prot exec map private map fixed map denywrite mmap prot read map private map fixed map denywrite mmap prot read prot write map private map fixed map denywrite close mmap null prot read prot write map private map anonymous mmap null prot read prot write map private map anonymous arch prctl arch set fs mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read mprotect prot read munmap set tid address set robust list rt sigaction sigrtmin sa handler sa mask sa flags sa restorer sa siginfo sa restorer null rt sigaction sigrt sa handler sa mask sa flags sa restorer sa restart sa siginfo sa restorer null rt sigprocmask sig unblock null rlimit stack null rlim cur rlim max infinity statfs sys fs selinux enoent no such file or directory statfs selinux enoent no such file or directory brk null brk openat at fdcwd proc filesystems o rdonly o cloexec fstat st mode s ifreg st size read nodev tsysfs nnodev ttmpfs nnodev tbd read close access etc selinux config f ok enoent no such file or directory openat at fdcwd usr lib locale locale archive o rdonly o cloexec fstat st mode s ifreg st size mmap null prot read map private close access proc self mounts r ok openat at fdcwd proc self mounts o rdonly o cloexec fstat st mode s ifreg st size read sysfs sys sysfs rw nosuid nodev lstat sys st mode s ifdir st size getcwd home cking lstat home cking sysfs enoent no such file or directory lstat proc st mode s ifdir st size getcwd home cking lstat home cking proc enoent no such file or directory lstat dev st mode s ifdir st size getcwd home cking lstat home cking udev enoent no such file or directory lstat dev st mode s ifdir st size lstat dev pts st mode s ifdir st size getcwd home cking lstat home cking devpts enoent no such file or directory lstat run st mode s ifdir st size getcwd home cking lstat home cking tmpfs enoent no such file or directory lstat dev st mode s ifdir st size lstat dev st mode s ifblk st rdev makedev stat st mode s ifdir st size statfs f type super magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid st relatime stat st mode s ifdir st size lstat sys st mode s ifdir st size lstat sys kernel st mode s ifdir st size lstat sys kernel security st mode s ifdir st size getcwd home cking lstat home cking securityfs enoent no such file or directory lstat dev st mode s ifdir st size lstat dev shm st mode s ifdir s isvtx st size getcwd home cking lstat home cking tmpfs enoent no such file or directory lstat run st mode s ifdir st size lstat run lock st mode s ifdir s isvtx st size getcwd home cking lstat home cking tmpfs enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size getcwd home cking lstat home cking tmpfs enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup unified st mode s ifdir st size getcwd home cking lstat home cking enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup systemd st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs pstore st mode s ifdir st size getcwd home cking lstat home cking pstore enoent no such file or directory lstat sys st mode s ifdir st size lstat sys firmware st mode s ifdir st size lstat sys firmware efi st mode s ifdir st size lstat sys firmware efi efivars st mode s ifdir st size getcwd home cking lstat home cking efivarfs enoent no such file or directory read xec relatime mode ncgroup lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs bpf st mode s ifdir s isvtx st size getcwd home cking lstat home cking none enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup cpu cpuacct st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup net cls net prio st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup rdma st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup devices st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup perf event st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup freezer st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup memory st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup pids st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup cpuset st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup hugetlb st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs cgroup st mode s ifdir st size lstat sys fs cgroup blkio st mode s ifdir st size getcwd home cking lstat home cking cgroup enoent no such file or directory read o direct pipe ino nhu lstat proc st mode s ifdir st size lstat proc sys st mode s ifdir st size lstat proc sys fs st mode s ifdir st size lstat proc sys fs binfmt misc st mode s ifdir st size getcwd home cking lstat home cking systemd enoent no such file or directory lstat dev st mode s ifdir st size lstat dev hugepages st mode s ifdir st size getcwd home cking lstat home cking hugetlbfs enoent no such file or directory lstat dev st mode s ifdir st size lstat dev mqueue st mode s ifdir s isvtx st size getcwd home cking lstat home cking mqueue enoent no such file or directory lstat sys st mode s ifdir st size lstat sys kernel st mode s ifdir st size lstat sys kernel debug st mode s ifdir st size getcwd home cking lstat home cking debugfs enoent no such file or directory lstat sys st mode s ifdir st size lstat sys kernel st mode s ifdir st size lstat sys kernel tracing st mode s ifdir st size getcwd home cking lstat home cking tracefs enoent no such file or directory lstat run st mode s ifdir st size lstat run rpc pipefs st mode s ifdir st size getcwd home cking lstat home cking sunrpc enoent no such file or directory lstat proc st mode s ifdir st size lstat proc fs st mode s ifdir st size lstat proc fs nfsd st mode s ifdir st size getcwd home cking lstat home cking nfsd enoent no such file or directory lstat sys st mode s ifdir st size lstat sys fs st mode s ifdir st size lstat sys fs fuse st mode s ifdir st size lstat sys fs fuse connections st mode s ifdir st size getcwd home cking lstat home cking fusectl enoent no such file or directory lstat sys st mode s ifdir st size lstat sys kernel st mode s ifdir st size lstat sys kernel config st mode s ifdir st size getcwd home cking lstat home cking configfs enoent no such file or directory lstat boot st mode s ifdir st size lstat boot efi st mode s ifdir st size lstat dev st mode s ifdir st size lstat dev st mode s ifblk st rdev makedev stat boot efi st mode s ifdir st size statfs boot efi f type msdos super magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid st relatime stat boot efi st mode s ifdir st size lstat snap st mode s ifdir st size lstat snap st mode s ifdir st size lstat snap st mode s ifdir st size lstat dev st mode s ifdir st size lstat dev st mode s ifblk st rdev makedev stat snap st mode s ifdir st size statfs snap f type squashfs magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid st rdonly st nodev st relatime stat snap st mode s ifdir st size lstat snap st mode s ifdir st size lstat snap st mode s ifdir st size lstat snap st mode s ifdir st size lstat dev st mode s ifdir st size lstat dev st mode s ifblk st rdev makedev stat snap st mode s ifdir st size statfs snap f type squashfs magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid st rdonly st nodev st relatime stat snap st mode s ifdir st size lstat snap st mode s ifdir st size lstat snap lxd st mode s ifdir st size lstat snap lxd st mode s ifdir st size lstat dev st mode s ifdir st size lstat dev st mode s ifblk st rdev makedev stat snap lxd st mode s ifdir st size statfs snap lxd f type squashfs magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid st rdonly st nodev st relatime stat snap lxd st mode s ifdir st size lstat snap st mode s ifdir st size lstat snap lxd st mode s ifdir st size lstat snap lxd st mode s ifdir st size lstat dev st mode s ifdir st size lstat dev st mode s ifblk st rdev makedev stat snap lxd st mode s ifdir st size statfs snap lxd f type squashfs magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid st rdonly st nodev st relatime stat snap lxd st mode s ifdir st size lstat snap st mode s ifdir st size lstat snap snapd st mode s ifdir st size lstat snap snapd st mode s ifdir st size lstat dev st mode s ifdir st size lstat dev st mode s ifblk st rdev makedev stat snap snapd st mode s ifdir st size statfs snap snapd f type squashfs magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid st rdonly st nodev st relatime stat snap snapd st mode s ifdir st size read mode nnsfs run lstat run st mode s ifdir st size lstat run snapd st mode s ifdir st size lstat run snapd ns st mode s ifdir st size getcwd home cking lstat home cking tmpfs enoent no such file or directory lstat run st mode s ifdir st size lstat run snapd st mode s ifdir st size lstat run snapd ns st mode s ifdir st size lstat run snapd ns lxd mnt st mode s ifreg st size getcwd home cking lstat home cking nsfs enoent no such file or directory lstat run st mode s ifdir st size lstat run user st mode s ifdir st size lstat run user st mode s ifdir st size getcwd home cking lstat home cking tmpfs enoent no such file or directory lstat testpool st mode s ifdir st size getcwd home cking lstat home cking testpool enoent no such file or directory lstat testpool st mode s ifdir st size lstat testpool test st mode s ifdir st size getcwd home cking lstat home cking testpool enoent no such file or directory read close openat at fdcwd etc projects o rdonly enoent no such file or directory stat testpool test dst enoent no such file or directory openat at fdcwd testpool test dst o rdwr o creat fstatfs f type zfs super magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid fstat st mode s ifreg st size fstatfs f type zfs super magic f bsize f blocks f bfree f bavail f files f ffree f fsid val f namelen f frsize f flags st valid fstat st mode s ifreg st size openat at fdcwd usr share locale locale alias o rdonly o cloexec fstat st mode s ifreg st size read locale name alias data base n read close openat at fdcwd usr share locale en gb utf lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en gb lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en gb lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en utf lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en gb utf lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en gb lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en gb lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en utf lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en lc messages xfsprogs mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en lc messages xfsprogs mo o rdonly enoent no such file or directory sysinfo uptime loads totalram freeram sharedram bufferram totalswap freeswap procs totalhigh freehigh mem unit stat testpool test src st mode s ifreg st size openat at fdcwd testpool test src o rdonly sendfile einval invalid argument dup fcntl f getfl flags o rdwr openat at fdcwd usr share locale en gb utf lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en gb lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en gb lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en utf lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale en lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en gb utf lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en gb lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en gb lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en utf lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en lc messages libc mo o rdonly enoent no such file or directory openat at fdcwd usr share locale langpack en lc messages libc mo o rdonly enoent no such file or directory fstat st mode s ifchr st rdev makedev write sendfile invalid argument n invalid argument close close fsync exit group exited with this works fine on with zfs | 1 |
8,926 | 2,612,931,200 | IssuesEvent | 2015-02-27 17:34:51 | chrsmith/windows-package-manager | https://api.github.com/repos/chrsmith/windows-package-manager | closed | Feature Request: Install a list of packages at once | auto-migrated Type-Defect | ```
One thing I love about linux package managers is, that you can install a whole
bunch of packages all at once. With Npackd it's easy to install many packages
in a very short time, but I think it would be great if you could give it a list
of packages to install.
This would also require that you can display a plain-text list of all packages
that were installed with npackd. For example, these are the packages I've
installed right now:
```
org.7-zip.SevenZIP64
org.7-zip.SevenZIPA
com.google.Chrome
org.eclipse.EclipseIDEJava64
com.github.bmatzelle.Gow
hashcheck
com.hdtune.hdtune
com.oracle.JavaAutoUpdater
com.oracle.JDK
com.oracle.JDK64
com.oracle.JRE
com.oracle.JRE64
meld
com.googlecode.msysgit.MSysGit
com.mysql.dev.MySQLWorkbench
com.googlecode.windows-package-manager.Npackd64
com.googlecode.windows-package-manager.NpackdInstallerHelper
com.googlecode.windows-package-manager.NpackdInstallerHelper
com.googlecode.windows-package-manager.NpackdCL
uk.org.greenend.chiark.sgtatham.Putty
org.python.Python
org.python.Python64
com.sourcetreeapp.SourceTree
com.teamviewer.complete
com.videolan.VLCMediaPlayer64
win32-disk-imager
net.winscp.WinSCP
```
I used `npackdcl.exe list --status=installed --bare-format` and then stripped
the package names. I also removed all packages beginning with msi.* because
npackd can't install them, and com.microsoft.* because these were libraries
that I either never installed actively or have been installed as a dependency.
There could be options similar to the settings-tab, where you can just copy-out
or paste-in a package-list. The generated list should only contain packages
which are installabe by Npackd.
I think it shouldn't be too much of a hassle to implement this and IMHO it
would a very useful feature. Especially for people like me, who reinstall
Windows twice a year.
This is just an idea of how I would do it (and maybe try to, if I find time). I
hope it doesn't sound too demanding.
Later, you could get the list via HTTP and BAM! Npackd can now manage thousands
of machines! :D
```
Original issue reported on code.google.com by `FUrba...@gmail.com` on 18 Jun 2014 at 3:16 | 1.0 | Feature Request: Install a list of packages at once - ```
One thing I love about linux package managers is, that you can install a whole
bunch of packages all at once. With Npackd it's easy to install many packages
in a very short time, but I think it would be great if you could give it a list
of packages to install.
This would also require that you can display a plain-text list of all packages
that were installed with npackd. For example, these are the packages I've
installed right now:
```
org.7-zip.SevenZIP64
org.7-zip.SevenZIPA
com.google.Chrome
org.eclipse.EclipseIDEJava64
com.github.bmatzelle.Gow
hashcheck
com.hdtune.hdtune
com.oracle.JavaAutoUpdater
com.oracle.JDK
com.oracle.JDK64
com.oracle.JRE
com.oracle.JRE64
meld
com.googlecode.msysgit.MSysGit
com.mysql.dev.MySQLWorkbench
com.googlecode.windows-package-manager.Npackd64
com.googlecode.windows-package-manager.NpackdInstallerHelper
com.googlecode.windows-package-manager.NpackdInstallerHelper
com.googlecode.windows-package-manager.NpackdCL
uk.org.greenend.chiark.sgtatham.Putty
org.python.Python
org.python.Python64
com.sourcetreeapp.SourceTree
com.teamviewer.complete
com.videolan.VLCMediaPlayer64
win32-disk-imager
net.winscp.WinSCP
```
I used `npackdcl.exe list --status=installed --bare-format` and then stripped
the package names. I also removed all packages beginning with msi.* because
npackd can't install them, and com.microsoft.* because these were libraries
that I either never installed actively or have been installed as a dependency.
There could be options similar to the settings-tab, where you can just copy-out
or paste-in a package-list. The generated list should only contain packages
which are installabe by Npackd.
I think it shouldn't be too much of a hassle to implement this and IMHO it
would a very useful feature. Especially for people like me, who reinstall
Windows twice a year.
This is just an idea of how I would do it (and maybe try to, if I find time). I
hope it doesn't sound too demanding.
Later, you could get the list via HTTP and BAM! Npackd can now manage thousands
of machines! :D
```
Original issue reported on code.google.com by `FUrba...@gmail.com` on 18 Jun 2014 at 3:16 | defect | feature request install a list of packages at once one thing i love about linux package managers is that you can install a whole bunch of packages all at once with npackd it s easy to install many packages in a very short time but i think it would be great if you could give it a list of packages to install this would also require that you can display a plain text list of all packages that were installed with npackd for example these are the packages i ve installed right now org zip org zip sevenzipa com google chrome org eclipse com github bmatzelle gow hashcheck com hdtune hdtune com oracle javaautoupdater com oracle jdk com oracle com oracle jre com oracle meld com googlecode msysgit msysgit com mysql dev mysqlworkbench com googlecode windows package manager com googlecode windows package manager npackdinstallerhelper com googlecode windows package manager npackdinstallerhelper com googlecode windows package manager npackdcl uk org greenend chiark sgtatham putty org python python org python com sourcetreeapp sourcetree com teamviewer complete com videolan disk imager net winscp winscp i used npackdcl exe list status installed bare format and then stripped the package names i also removed all packages beginning with msi because npackd can t install them and com microsoft because these were libraries that i either never installed actively or have been installed as a dependency there could be options similar to the settings tab where you can just copy out or paste in a package list the generated list should only contain packages which are installabe by npackd i think it shouldn t be too much of a hassle to implement this and imho it would a very useful feature especially for people like me who reinstall windows twice a year this is just an idea of how i would do it and maybe try to if i find time i hope it doesn t sound too demanding later you could get the list via http and bam npackd can now manage thousands of machines d original issue reported on code google com by furba gmail com on jun at | 1 |
16,401 | 2,891,834,661 | IssuesEvent | 2015-06-15 08:58:56 | danieleteti/delphi-orm | https://api.github.com/repos/danieleteti/delphi-orm | closed | Optimistic concurrency control | auto-migrated Priority-Medium Type-Defect | ```
Just a simple question
Have you provided an optimistic concurrency control for object persistence.
Hibernate has it through @Optimistic annotation on a field.
It is basically a date/int field for each table/object in database and gets
compared/incremented at the end of each transaction commit.
```
Original issue reported on code.google.com by `Sergei.I...@gmail.com` on 4 Mar 2015 at 9:11 | 1.0 | Optimistic concurrency control - ```
Just a simple question
Have you provided an optimistic concurrency control for object persistence.
Hibernate has it through @Optimistic annotation on a field.
It is basically a date/int field for each table/object in database and gets
compared/incremented at the end of each transaction commit.
```
Original issue reported on code.google.com by `Sergei.I...@gmail.com` on 4 Mar 2015 at 9:11 | defect | optimistic concurrency control just a simple question have you provided an optimistic concurrency control for object persistence hibernate has it through optimistic annotation on a field it is basically a date int field for each table object in database and gets compared incremented at the end of each transaction commit original issue reported on code google com by sergei i gmail com on mar at | 1 |
418,131 | 12,193,767,689 | IssuesEvent | 2020-04-29 14:52:03 | GingerWalnut/SQBeyondPublic | https://api.github.com/repos/GingerWalnut/SQBeyondPublic | closed | Weird issue on Izifo | bug medium priority | Ok so I went to Izifo earlier today and got kicked. Kept trying to relog and always got the java exception about connection forced closed by host. I went on the discord looking for help and Tiat was able to fly me and my ship back into space. Now everything works as usual. Fixed just like that. Im not sure what happened but what i think was my connection. I have rural internet and it fluctuates between 10-20mbps. That and maybe all the trees on Izifo lagged me out possibly. Seems this was totally client side. And we thought it would be worth mentioning. On a side not im staying far away from Izifo lol.
Edit: Also would like to add it resulted in a kick from minechat app as well. Dont know if that helps. | 1.0 | Weird issue on Izifo - Ok so I went to Izifo earlier today and got kicked. Kept trying to relog and always got the java exception about connection forced closed by host. I went on the discord looking for help and Tiat was able to fly me and my ship back into space. Now everything works as usual. Fixed just like that. Im not sure what happened but what i think was my connection. I have rural internet and it fluctuates between 10-20mbps. That and maybe all the trees on Izifo lagged me out possibly. Seems this was totally client side. And we thought it would be worth mentioning. On a side not im staying far away from Izifo lol.
Edit: Also would like to add it resulted in a kick from minechat app as well. Dont know if that helps. | non_defect | weird issue on izifo ok so i went to izifo earlier today and got kicked kept trying to relog and always got the java exception about connection forced closed by host i went on the discord looking for help and tiat was able to fly me and my ship back into space now everything works as usual fixed just like that im not sure what happened but what i think was my connection i have rural internet and it fluctuates between that and maybe all the trees on izifo lagged me out possibly seems this was totally client side and we thought it would be worth mentioning on a side not im staying far away from izifo lol edit also would like to add it resulted in a kick from minechat app as well dont know if that helps | 0 |
40,983 | 10,256,099,834 | IssuesEvent | 2019-08-21 16:50:45 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | [2.x] CookieComponent problems | component defect | This is a bug.
* CakePHP Version: 2.10.3
* Platform and Target: Non-related
This ticket contains multiple problems of 2.x CookieComponent. I am planning to send multiple pull requests to fix them step by step.
### Problems
* [x] Problem 1 - The write() method throws fatal error when replacing scalar with array
* [ ] Problem 2 - The read() method restores original values if cookies have been completely deleted
* [x] Problem 3 - Deleting deep entry doesn't work (As of 2.4.3)
* [ ] Problem 4 - Writing cookie with expiration time for specific entry doesn't work (As of 2.4.3)
### TestsController
```php
class TestsController extends AppController
{
public $autoRender = false;
public function read() {
debug( $this->Cookie->read() );
}
public function write() {
$this->Cookie->write('foo.bar', 'value');
}
public function deleteDeepEntry() {
$this->Cookie->delete('foo.bar');
}
public function replaceScalarWithArray() {
$this->Cookie->write('foo', 1);
$this->Cookie->write('foo.bar', 2);
}
public function readAfterDelete() {
$this->Cookie->delete('foo');
debug( $this->Cookie->read() );
}
public function writeWithExpirationTime() {
$this->Cookie->write('foo.baz', 'value', true, '5 second');
}
}
```
----
### Problem 1
The write() method throws fatal error when replacing scalar with array.
#### Test
1. Access /tests/replaceScalarWithArray
#### Expected
Works.
#### Actual
```
Argument 1 passed to Hash::insert() must be of the type array, integer given, called in ...
```
----
### Problem 2
The read() method restores original values if cookies have been completely deleted.
#### Test
1. Access /tests/write
2. Access /tests/readAfterDelete
#### Expected
```
array()
```
#### Actual
```
array(
'foo' => array(
'bar' => 'value'
)
)
```
----
### Problem 3
Deleting deep entry doesn't work (As of 2.4.3).
#### Test
1. Access /tests/write
2. Access /tests/deleteDeepEntry
3. Access /tests/read
#### Expected
```
array()
```
#### Actual
```
array(
'foo' => array(
'bar' => 'value'
)
)
```
----
### Problem 4
Writing cookie with expiration time for specific entry doesn't work (As of 2.4.3).
#### Test
1. Access /tests/write
2. Access /tests/writeWithExpirationTime
3. Access /tests/read (3 seconds later)
3. Access /tests/read (5 seconds later)
#### Expected
(3 seconds later)
```
array(
'foo' => array(
'bar' => 'value',
'baz' => 'value'
)
)
```
(5 seconds later)
```
array(
'foo' => array(
'bar' => 'value',
)
)
```
#### Actual
(3 seconds later)
```
array(
'foo' => array(
'bar' => 'value',
'baz' => 'value'
)
)
```
(5 seconds later)
```
array()
```
----
### Related tickets
#10341 | 1.0 | [2.x] CookieComponent problems - This is a bug.
* CakePHP Version: 2.10.3
* Platform and Target: Non-related
This ticket contains multiple problems of 2.x CookieComponent. I am planning to send multiple pull requests to fix them step by step.
### Problems
* [x] Problem 1 - The write() method throws fatal error when replacing scalar with array
* [ ] Problem 2 - The read() method restores original values if cookies have been completely deleted
* [x] Problem 3 - Deleting deep entry doesn't work (As of 2.4.3)
* [ ] Problem 4 - Writing cookie with expiration time for specific entry doesn't work (As of 2.4.3)
### TestsController
```php
class TestsController extends AppController
{
public $autoRender = false;
public function read() {
debug( $this->Cookie->read() );
}
public function write() {
$this->Cookie->write('foo.bar', 'value');
}
public function deleteDeepEntry() {
$this->Cookie->delete('foo.bar');
}
public function replaceScalarWithArray() {
$this->Cookie->write('foo', 1);
$this->Cookie->write('foo.bar', 2);
}
public function readAfterDelete() {
$this->Cookie->delete('foo');
debug( $this->Cookie->read() );
}
public function writeWithExpirationTime() {
$this->Cookie->write('foo.baz', 'value', true, '5 second');
}
}
```
----
### Problem 1
The write() method throws fatal error when replacing scalar with array.
#### Test
1. Access /tests/replaceScalarWithArray
#### Expected
Works.
#### Actual
```
Argument 1 passed to Hash::insert() must be of the type array, integer given, called in ...
```
----
### Problem 2
The read() method restores original values if cookies have been completely deleted.
#### Test
1. Access /tests/write
2. Access /tests/readAfterDelete
#### Expected
```
array()
```
#### Actual
```
array(
'foo' => array(
'bar' => 'value'
)
)
```
----
### Problem 3
Deleting deep entry doesn't work (As of 2.4.3).
#### Test
1. Access /tests/write
2. Access /tests/deleteDeepEntry
3. Access /tests/read
#### Expected
```
array()
```
#### Actual
```
array(
'foo' => array(
'bar' => 'value'
)
)
```
----
### Problem 4
Writing cookie with expiration time for specific entry doesn't work (As of 2.4.3).
#### Test
1. Access /tests/write
2. Access /tests/writeWithExpirationTime
3. Access /tests/read (3 seconds later)
3. Access /tests/read (5 seconds later)
#### Expected
(3 seconds later)
```
array(
'foo' => array(
'bar' => 'value',
'baz' => 'value'
)
)
```
(5 seconds later)
```
array(
'foo' => array(
'bar' => 'value',
)
)
```
#### Actual
(3 seconds later)
```
array(
'foo' => array(
'bar' => 'value',
'baz' => 'value'
)
)
```
(5 seconds later)
```
array()
```
----
### Related tickets
#10341 | defect | cookiecomponent problems this is a bug cakephp version platform and target non related this ticket contains multiple problems of x cookiecomponent i am planning to send multiple pull requests to fix them step by step problems problem the write method throws fatal error when replacing scalar with array problem the read method restores original values if cookies have been completely deleted problem deleting deep entry doesn t work as of problem writing cookie with expiration time for specific entry doesn t work as of testscontroller php class testscontroller extends appcontroller public autorender false public function read debug this cookie read public function write this cookie write foo bar value public function deletedeepentry this cookie delete foo bar public function replacescalarwitharray this cookie write foo this cookie write foo bar public function readafterdelete this cookie delete foo debug this cookie read public function writewithexpirationtime this cookie write foo baz value true second problem the write method throws fatal error when replacing scalar with array test access tests replacescalarwitharray expected works actual argument passed to hash insert must be of the type array integer given called in problem the read method restores original values if cookies have been completely deleted test access tests write access tests readafterdelete expected array actual array foo array bar value problem deleting deep entry doesn t work as of test access tests write access tests deletedeepentry access tests read expected array actual array foo array bar value problem writing cookie with expiration time for specific entry doesn t work as of test access tests write access tests writewithexpirationtime access tests read seconds later access tests read seconds later expected seconds later array foo array bar value baz value seconds later array foo array bar value actual seconds later array foo array bar value baz value seconds later array related tickets | 1 |
71,300 | 23,529,544,066 | IssuesEvent | 2022-08-19 14:05:57 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | [Spaces] Something wrong happens when I join a room listed in a sub-space without joining the sub-space before | T-Defect X-Needs-Info | ### Steps to reproduce
I reproduced it several time
- explore the rooms of a joined space
- decide to join a room listed under a sub-space whereas the sub-space is not joined yet
- the room is joined, but Element desktop prompts the user something wrong happened and suggests reporting a bug
### Outcome
the joined room should be displayed
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | [Spaces] Something wrong happens when I join a room listed in a sub-space without joining the sub-space before - ### Steps to reproduce
I reproduced it several time
- explore the rooms of a joined space
- decide to join a room listed under a sub-space whereas the sub-space is not joined yet
- the room is joined, but Element desktop prompts the user something wrong happened and suggests reporting a bug
### Outcome
the joined room should be displayed
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Will you send logs?
No | defect | something wrong happens when i join a room listed in a sub space without joining the sub space before steps to reproduce i reproduced it several time explore the rooms of a joined space decide to join a room listed under a sub space whereas the sub space is not joined yet the room is joined but element desktop prompts the user something wrong happened and suggests reporting a bug outcome the joined room should be displayed operating system no response application version no response how did you install the app no response homeserver no response will you send logs no | 1 |
51,346 | 13,207,439,614 | IssuesEvent | 2020-08-14 23:06:25 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | Python documentation (Trac #260) | Incomplete Migration Migrated from Trac combo core defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/260">https://code.icecube.wisc.edu/projects/icecube/ticket/260</a>, reported by nega</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-11T21:19:20",
"_ts": "1423689560987002",
"description": "Ensure that all python code has proper documentation.",
"reporter": "nega",
"cc": "blaufuss, olivas",
"resolution": "wontfix",
"time": "2011-05-11T20:32:37",
"component": "combo core",
"summary": "Python documentation",
"priority": "normal",
"keywords": "python documentation",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Python documentation (Trac #260) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/260">https://code.icecube.wisc.edu/projects/icecube/ticket/260</a>, reported by nega</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-11T21:19:20",
"_ts": "1423689560987002",
"description": "Ensure that all python code has proper documentation.",
"reporter": "nega",
"cc": "blaufuss, olivas",
"resolution": "wontfix",
"time": "2011-05-11T20:32:37",
"component": "combo core",
"summary": "Python documentation",
"priority": "normal",
"keywords": "python documentation",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | python documentation trac migrated from json status closed changetime ts description ensure that all python code has proper documentation reporter nega cc blaufuss olivas resolution wontfix time component combo core summary python documentation priority normal keywords python documentation milestone owner type defect | 1 |
485,410 | 13,965,092,206 | IssuesEvent | 2020-10-25 20:54:39 | getodk/collect | https://api.github.com/repos/getodk/collect | closed | Constraints not immediately evaluated correctly when editing saved form | bug high priority | #### Software and hardware versions
Collect v1.28
#### Problem description
Constraints not immediately evaluated correctly when editing saved form.
#### Steps to reproduce the problem
1. Load any form with a constraint
1. Fill out some data, save it
1. Edit the saved record
1. Put in an invalid value and notice you can swipe forward
1. Swipe back twice and notice you see the constraint message
1. Put in a valid value and notice you can't swipe forward
1. Swipe back twice and notice you can swipe forward (the valid value is accepted)
#### Expected behavior
Constraints should work correctly on the first try even when editing an existing instance.
#### Other information
This is related to https://github.com/getodk/javarosa/pull/578. The main instance name is being set on instance deserialization which makes certain `TreeReference` comparisons fail including the one to check whether the user input should be used instead of the model value. | 1.0 | Constraints not immediately evaluated correctly when editing saved form - #### Software and hardware versions
Collect v1.28
#### Problem description
Constraints not immediately evaluated correctly when editing saved form.
#### Steps to reproduce the problem
1. Load any form with a constraint
1. Fill out some data, save it
1. Edit the saved record
1. Put in an invalid value and notice you can swipe forward
1. Swipe back twice and notice you see the constraint message
1. Put in a valid value and notice you can't swipe forward
1. Swipe back twice and notice you can swipe forward (the valid value is accepted)
#### Expected behavior
Constraints should work correctly on the first try even when editing an existing instance.
#### Other information
This is related to https://github.com/getodk/javarosa/pull/578. The main instance name is being set on instance deserialization which makes certain `TreeReference` comparisons fail including the one to check whether the user input should be used instead of the model value. | non_defect | constraints not immediately evaluated correctly when editing saved form software and hardware versions collect problem description constraints not immediately evaluated correctly when editing saved form steps to reproduce the problem load any form with a constraint fill out some data save it edit the saved record put in an invalid value and notice you can swipe forward swipe back twice and notice you see the constraint message put in a valid value and notice you can t swipe forward swipe back twice and notice you can swipe forward the valid value is accepted expected behavior constraints should work correctly on the first try even when editing an existing instance other information this is related to the main instance name is being set on instance deserialization which makes certain treereference comparisons fail including the one to check whether the user input should be used instead of the model value | 0 |
148,403 | 5,681,230,903 | IssuesEvent | 2017-04-13 05:25:07 | psouza4/mediacentermaster | https://api.github.com/repos/psouza4/mediacentermaster | closed | Usenet Manager window crash/exception error | Affects-Stability Component-Functionality Feature-Downloads Feature-Usenet Fixed / Resolved Priority-High Type-Bug | As detailed:
http://forums.mediacentermaster.com/viewtopic.php?f=6&t=11565
http://forums.mediacentermaster.com/viewtopic.php?f=4&t=11829
And other topics.
The exception occurs when adding new downloads in a public build version of Media Center Master. | 1.0 | Usenet Manager window crash/exception error - As detailed:
http://forums.mediacentermaster.com/viewtopic.php?f=6&t=11565
http://forums.mediacentermaster.com/viewtopic.php?f=4&t=11829
And other topics.
The exception occurs when adding new downloads in a public build version of Media Center Master. | non_defect | usenet manager window crash exception error as detailed and other topics the exception occurs when adding new downloads in a public build version of media center master | 0 |
7,344 | 2,610,364,389 | IssuesEvent | 2015-02-26 19:57:46 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | Not working | auto-migrated Priority-Medium Type-Defect | ```
Can not post and it does not show editor for changing fonts and other features
What browser are you using? FIREFOX
What version of ScribeFire are you running? ScireFire 4
```
-----
Original issue reported on code.google.com by `LUVNUR...@gmail.com` on 23 Oct 2012 at 2:31 | 1.0 | Not working - ```
Can not post and it does not show editor for changing fonts and other features
What browser are you using? FIREFOX
What version of ScribeFire are you running? ScireFire 4
```
-----
Original issue reported on code.google.com by `LUVNUR...@gmail.com` on 23 Oct 2012 at 2:31 | defect | not working can not post and it does not show editor for changing fonts and other features what browser are you using firefox what version of scribefire are you running scirefire original issue reported on code google com by luvnur gmail com on oct at | 1 |
167,422 | 20,726,070,939 | IssuesEvent | 2022-03-14 02:08:57 | t2kx/juice-shop | https://api.github.com/repos/t2kx/juice-shop | opened | CVE-2021-37701 (High) detected in tar-2.2.2.tgz | security vulnerability | ## CVE-2021-37701 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.2.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.2.tgz">https://registry.npmjs.org/tar/-/tar-2.2.2.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/node-gyp/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.14.0.tgz (Root Library)
- node-gyp-3.8.0.tgz
- :x: **tar-2.2.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution (tar): 4.4.16</p>
<p>Direct dependency fix Resolution (node-sass): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-37701 (High) detected in tar-2.2.2.tgz - ## CVE-2021-37701 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.2.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.2.tgz">https://registry.npmjs.org/tar/-/tar-2.2.2.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/node-gyp/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.14.0.tgz (Root Library)
- node-gyp-3.8.0.tgz
- :x: **tar-2.2.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution (tar): 4.4.16</p>
<p>Direct dependency fix Resolution (node-sass): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules node gyp node modules tar package json dependency hierarchy node sass tgz root library node gyp tgz x tar tgz vulnerable library found in base branch master vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution node sass step up your open source security game with whitesource | 0 |
24,610 | 4,044,992,909 | IssuesEvent | 2016-05-21 17:56:13 | TASVideos/BizHawk | https://api.github.com/repos/TASVideos/BizHawk | closed | Graphical glitch in Atari 2600 Crystal Castles | auto-migrated Core-Atari2600Hawk Priority-Low Type-Defect | ```
What steps will reproduce the problem?
1. Start the game
2.
3.
What is the expected output? What do you see instead?
Expected: game draws the first castle. See second attachment for Stella
emulation.
Seen: game draws first castle, there is also garbage lines in upper-left. See
first attachment for BizHawk emulation.
What version of the product are you using? On what operating system?
1.8.1, Windows 7 64-bit
Please provide any additional information below.
I do not believe this graphical glitch affects gameplay.
```
Original issue reported on code.google.com by `djk...@gmail.com` on 31 Aug 2014 at 4:29
Attachments:
* [Crystal Castles (1984) (Atari).2014-08-30 21.23.20.png](https://storage.googleapis.com/google-code-attachments/bizhawk/issue-260/comment-0/Crystal Castles (1984) (Atari).2014-08-30 21.23.20.png)
* [Crystal Castles (1984) (Atari).png](https://storage.googleapis.com/google-code-attachments/bizhawk/issue-260/comment-0/Crystal Castles (1984) (Atari).png)
| 1.0 | Graphical glitch in Atari 2600 Crystal Castles - ```
What steps will reproduce the problem?
1. Start the game
2.
3.
What is the expected output? What do you see instead?
Expected: game draws the first castle. See second attachment for Stella
emulation.
Seen: game draws first castle, there is also garbage lines in upper-left. See
first attachment for BizHawk emulation.
What version of the product are you using? On what operating system?
1.8.1, Windows 7 64-bit
Please provide any additional information below.
I do not believe this graphical glitch affects gameplay.
```
Original issue reported on code.google.com by `djk...@gmail.com` on 31 Aug 2014 at 4:29
Attachments:
* [Crystal Castles (1984) (Atari).2014-08-30 21.23.20.png](https://storage.googleapis.com/google-code-attachments/bizhawk/issue-260/comment-0/Crystal Castles (1984) (Atari).2014-08-30 21.23.20.png)
* [Crystal Castles (1984) (Atari).png](https://storage.googleapis.com/google-code-attachments/bizhawk/issue-260/comment-0/Crystal Castles (1984) (Atari).png)
| defect | graphical glitch in atari crystal castles what steps will reproduce the problem start the game what is the expected output what do you see instead expected game draws the first castle see second attachment for stella emulation seen game draws first castle there is also garbage lines in upper left see first attachment for bizhawk emulation what version of the product are you using on what operating system windows bit please provide any additional information below i do not believe this graphical glitch affects gameplay original issue reported on code google com by djk gmail com on aug at attachments castles atari png castles atari png | 1 |
76,977 | 26,704,350,339 | IssuesEvent | 2023-01-27 16:50:04 | bigbluebutton/bigbluebutton | https://api.github.com/repos/bigbluebutton/bigbluebutton | closed | The loading of the PDF is interrupted | type: defect priority: normal module: client | **Describe the bug**
The loading of "larger" PDFs is interrupted if you connect the camera fast enough or open the browser console, for example.
```
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <path> attribute d: Expected number, "M -Infinity,-Infin…".
w @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Js @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
au @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ou @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:416
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o.withComputation @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_compute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_recompute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o._runFlush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
a @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
postMessage (async)
u @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
v @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
invalidate @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
changed @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
t.runTask @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.flush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.drain @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:400
t @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
Meteor.connection._processOneDataMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
_livedata_data @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
onMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
forEachCallback @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
socket.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
r.dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._dispatchMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._didMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
o.ws.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <path> attribute d: Expected number, "M NaN,NaN L NaN,Na…".
w @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Js @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
au @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ou @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:283
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
e @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
setTimeout (async)
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
du @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
lu @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:416
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o.withComputation @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_compute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_recompute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o._runFlush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
a @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
postMessage (async)
u @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
v @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
invalidate @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
changed @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
t.runTask @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.flush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.drain @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:400
t @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
Meteor.connection._processOneDataMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
_livedata_data @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
onMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
forEachCallback @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
socket.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
r.dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._dispatchMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._didMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
o.ws.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <g> attribute transform: Expected number, "scale(NaN) translate(N…".
w @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Js @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
au @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ou @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:283
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
e @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
setTimeout (async)
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
du @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
lu @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:416
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o.withComputation @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_compute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_recompute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o._runFlush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
a @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
postMessage (async)
u @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
v @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
invalidate @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
changed @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
t.runTask @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.flush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.drain @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:400
t @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
Meteor.connection._processOneDataMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
_livedata_data @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
onMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
forEachCallback @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
socket.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
r.dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._dispatchMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._didMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
o.ws.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <g> attribute transform: Expected number, "scale(Infinity) transl…".
```
**To Reproduce**
Steps to reproduce the behavior:
1. Start a meeting an upload a larger PDF (10MB for example)
2. Connect the audio as quickly as possible
3. Connect the camera as quickly as possible or open the browser console
4. The pdf ist not loaded
**Screenshots**


**BBB version:**
BigBlueButton Server 2.6.0-beta.7 (759)
**Desktop (please complete the following information):**
- latest chrome
| 1.0 | The loading of the PDF is interrupted - **Describe the bug**
The loading of "larger" PDFs is interrupted if you connect the camera fast enough or open the browser console, for example.
```
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <path> attribute d: Expected number, "M -Infinity,-Infin…".
w @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Js @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
au @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ou @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:416
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o.withComputation @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_compute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_recompute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o._runFlush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
a @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
postMessage (async)
u @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
v @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
invalidate @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
changed @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
t.runTask @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.flush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.drain @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:400
t @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
Meteor.connection._processOneDataMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
_livedata_data @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
onMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
forEachCallback @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
socket.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
r.dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._dispatchMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._didMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
o.ws.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <path> attribute d: Expected number, "M NaN,NaN L NaN,Na…".
w @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Js @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
au @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ou @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:283
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
e @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
setTimeout (async)
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
du @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
lu @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:416
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o.withComputation @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_compute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_recompute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o._runFlush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
a @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
postMessage (async)
u @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
v @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
invalidate @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
changed @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
t.runTask @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.flush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.drain @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:400
t @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
Meteor.connection._processOneDataMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
_livedata_data @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
onMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
forEachCallback @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
socket.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
r.dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._dispatchMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._didMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
o.ws.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <g> attribute transform: Expected number, "scale(NaN) translate(N…".
w @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Js @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
au @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ou @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:283
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
e @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:253
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:9
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
setTimeout (async)
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
du @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
lu @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Ul @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
n.unstable_runWithPriority @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:79
bo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Fo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
yo @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
Hl @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
ns @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:416
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o.withComputation @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_compute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
_recompute @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
o._runFlush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
a @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
postMessage (async)
u @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
v @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
invalidate @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
changed @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:374
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
t.runTask @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.flush @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
t.drain @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:1
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:376
insert @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:400
t @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
Meteor.connection._processOneDataMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:434
_livedata_data @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
onMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:390
(anonymous) @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
forEachCallback @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
socket.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
r.dispatchEvent @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._dispatchMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
k._didMessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
o.ws.onmessage @ 42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:388
42b9b622b198b0af7320aa69258d95d1f79bc6b5.js?meteor_js_resource=true:61 Error: <g> attribute transform: Expected number, "scale(Infinity) transl…".
```
**To Reproduce**
Steps to reproduce the behavior:
1. Start a meeting an upload a larger PDF (10MB for example)
2. Connect the audio as quickly as possible
3. Connect the camera as quickly as possible or open the browser console
4. The pdf ist not loaded
**Screenshots**


**BBB version:**
BigBlueButton Server 2.6.0-beta.7 (759)
**Desktop (please complete the following information):**
- latest chrome
| defect | the loading of the pdf is interrupted describe the bug the loading of larger pdfs is interrupted if you connect the camera fast enough or open the browser console for example js meteor js resource true error attribute d expected number m infinity infin… w js meteor js resource true js js meteor js resource true au js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true ou js meteor js resource true ul js meteor js resource true anonymous js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true fo js meteor js resource true yo js meteor js resource true hl js meteor js resource true ns js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true o withcomputation js meteor js resource true compute js meteor js resource true recompute js meteor js resource true o runflush js meteor js resource true a js meteor js resource true postmessage async u js meteor js resource true v js meteor js resource true invalidate js meteor js resource true changed js meteor js resource true anonymous js meteor js resource true t runtask js meteor js resource true t flush js meteor js resource true t drain js meteor js resource true insert js meteor js resource true insert js meteor js resource true t js meteor js resource true dispatchevent js meteor js resource true meteor connection processonedatamessage js meteor js resource true livedata data js meteor js resource true onmessage js meteor js resource true anonymous js meteor js resource true foreachcallback js meteor js resource true socket onmessage js meteor js resource true r dispatchevent js meteor js resource true k dispatchmessage js meteor js resource true k didmessage js meteor js resource true o ws onmessage js meteor js resource true js meteor js resource true error attribute d expected number m nan nan l nan na… w js meteor js resource true js js meteor js resource true au js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true ou js meteor js resource true ul js meteor js resource true anonymous js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true fo js meteor js resource true yo js meteor js resource true hl js meteor js resource true ns js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true e js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true settimeout async anonymous js meteor js resource true du js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true lu js meteor js resource true ul js meteor js resource true anonymous js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true fo js meteor js resource true yo js meteor js resource true hl js meteor js resource true ns js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true o withcomputation js meteor js resource true compute js meteor js resource true recompute js meteor js resource true o runflush js meteor js resource true a js meteor js resource true postmessage async u js meteor js resource true v js meteor js resource true invalidate js meteor js resource true changed js meteor js resource true anonymous js meteor js resource true t runtask js meteor js resource true t flush js meteor js resource true t drain js meteor js resource true insert js meteor js resource true insert js meteor js resource true t js meteor js resource true dispatchevent js meteor js resource true meteor connection processonedatamessage js meteor js resource true livedata data js meteor js resource true onmessage js meteor js resource true anonymous js meteor js resource true foreachcallback js meteor js resource true socket onmessage js meteor js resource true r dispatchevent js meteor js resource true k dispatchmessage js meteor js resource true k didmessage js meteor js resource true o ws onmessage js meteor js resource true js meteor js resource true error attribute transform expected number scale nan translate n… w js meteor js resource true js js meteor js resource true au js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true ou js meteor js resource true ul js meteor js resource true anonymous js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true fo js meteor js resource true yo js meteor js resource true hl js meteor js resource true ns js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true e js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true settimeout async anonymous js meteor js resource true du js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true lu js meteor js resource true ul js meteor js resource true anonymous js meteor js resource true n unstable runwithpriority js meteor js resource true bo js meteor js resource true fo js meteor js resource true yo js meteor js resource true hl js meteor js resource true ns js meteor js resource true anonymous js meteor js resource true anonymous js meteor js resource true o withcomputation js meteor js resource true compute js meteor js resource true recompute js meteor js resource true o runflush js meteor js resource true a js meteor js resource true postmessage async u js meteor js resource true v js meteor js resource true invalidate js meteor js resource true changed js meteor js resource true anonymous js meteor js resource true t runtask js meteor js resource true t flush js meteor js resource true t drain js meteor js resource true insert js meteor js resource true insert js meteor js resource true t js meteor js resource true dispatchevent js meteor js resource true meteor connection processonedatamessage js meteor js resource true livedata data js meteor js resource true onmessage js meteor js resource true anonymous js meteor js resource true foreachcallback js meteor js resource true socket onmessage js meteor js resource true r dispatchevent js meteor js resource true k dispatchmessage js meteor js resource true k didmessage js meteor js resource true o ws onmessage js meteor js resource true js meteor js resource true error attribute transform expected number scale infinity transl… to reproduce steps to reproduce the behavior start a meeting an upload a larger pdf for example connect the audio as quickly as possible connect the camera as quickly as possible or open the browser console the pdf ist not loaded screenshots bbb version bigbluebutton server beta desktop please complete the following information latest chrome | 1 |
404,545 | 27,490,087,845 | IssuesEvent | 2023-03-04 14:09:35 | jocosocial/swiftarr | https://api.github.com/repos/jocosocial/swiftarr | closed | 2023 How-To Guide Enhancements | documentation 2022 | - [ ] Time Zone Check page
- [ ] Map (maybe include this? idk)
- [ ] Open Chats
- [ ] Search (it moved, maybe a bit more visible)
- [ ] Board game guide
- [ ] FAQ page (on login and home) | 1.0 | 2023 How-To Guide Enhancements - - [ ] Time Zone Check page
- [ ] Map (maybe include this? idk)
- [ ] Open Chats
- [ ] Search (it moved, maybe a bit more visible)
- [ ] Board game guide
- [ ] FAQ page (on login and home) | non_defect | how to guide enhancements time zone check page map maybe include this idk open chats search it moved maybe a bit more visible board game guide faq page on login and home | 0 |
35,946 | 9,691,016,116 | IssuesEvent | 2019-05-24 10:02:11 | Lundalogik/lip | https://api.github.com/repos/Lundalogik/lip | opened | Support for Chromium | bug package builder | The package builder does not work with Chromium. This needs to be fixed since the desktop client will move completely to chromium within a few releases.
Tricky situation: We want the package builder to work also for customers not using Chromium yet. Or can we say that once we fix support for Chromium we will release a new major version and customers using older DC will not be able to use the package builder? | 1.0 | Support for Chromium - The package builder does not work with Chromium. This needs to be fixed since the desktop client will move completely to chromium within a few releases.
Tricky situation: We want the package builder to work also for customers not using Chromium yet. Or can we say that once we fix support for Chromium we will release a new major version and customers using older DC will not be able to use the package builder? | non_defect | support for chromium the package builder does not work with chromium this needs to be fixed since the desktop client will move completely to chromium within a few releases tricky situation we want the package builder to work also for customers not using chromium yet or can we say that once we fix support for chromium we will release a new major version and customers using older dc will not be able to use the package builder | 0 |
61,578 | 17,023,729,809 | IssuesEvent | 2021-07-03 03:31:55 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | [landcover] landuse=military covers leisure=pitch | Component: mapnik Priority: minor Resolution: wontfix Type: defect | **[Submitted to the original trac issue database at 9.57pm, Monday, 4th July 2011]**
the (ugly) lines from landuse=military cover areas with leisure=pitch, sports=x.
see: http://osm.org/go/0JhLXApnD- | 1.0 | [landcover] landuse=military covers leisure=pitch - **[Submitted to the original trac issue database at 9.57pm, Monday, 4th July 2011]**
the (ugly) lines from landuse=military cover areas with leisure=pitch, sports=x.
see: http://osm.org/go/0JhLXApnD- | defect | landuse military covers leisure pitch the ugly lines from landuse military cover areas with leisure pitch sports x see | 1 |
7,663 | 25,395,021,426 | IssuesEvent | 2022-11-22 07:58:29 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | opened | When OCP is upgraded, wait for the catalog source to become ready on UI | bug ui_automation | test- tests/ui/test_validation_ui.py::TestUserInterfaceValidation::test_dashboard_validation_ui
ocs-ci results for OCS4-7-Downstream-OCP4-8-AWS-IPI-3AZ-RHCOS-3M-3W-tier1-or-tier_after_upgrade-post-upgrade (BUILD ID: 4.7.14-268.ci RUN ID: 1667455542)
The test failed with
```
Message: AssertionError: ['Healthy string not found on subscription_tab']
Type: None
```
This is because we are upgrading OCP, and we should wait for the catalog source to become healthy on UI before failing the test.
A direct assertion will immediately fail the test just like in this case. | 1.0 | When OCP is upgraded, wait for the catalog source to become ready on UI - test- tests/ui/test_validation_ui.py::TestUserInterfaceValidation::test_dashboard_validation_ui
ocs-ci results for OCS4-7-Downstream-OCP4-8-AWS-IPI-3AZ-RHCOS-3M-3W-tier1-or-tier_after_upgrade-post-upgrade (BUILD ID: 4.7.14-268.ci RUN ID: 1667455542)
The test failed with
```
Message: AssertionError: ['Healthy string not found on subscription_tab']
Type: None
```
This is because we are upgrading OCP, and we should wait for the catalog source to become healthy on UI before failing the test.
A direct assertion will immediately fail the test just like in this case. | non_defect | when ocp is upgraded wait for the catalog source to become ready on ui test tests ui test validation ui py testuserinterfacevalidation test dashboard validation ui ocs ci results for downstream aws ipi rhcos or tier after upgrade post upgrade build id ci run id the test failed with message assertionerror type none this is because we are upgrading ocp and we should wait for the catalog source to become healthy on ui before failing the test a direct assertion will immediately fail the test just like in this case | 0 |
26,070 | 4,567,957,743 | IssuesEvent | 2016-09-15 13:05:48 | jtomass/spring2 | https://api.github.com/repos/jtomass/spring2 | closed | SpringUIProvider's webApplicationContext field is not serializable | Defect Imported from Trac | **Originally by _johannes.tuikkala_ on 2016-01-13 09:02:40 +0200**
___
```
org.springframework.data.redis.serializer.SerializationException: Cannot serialize; nested exception is org.springframework.core.serializer.support.SerializationFailedException: Failed to serialize object using DefaultSerializer; nested exception is java.io.NotSerializableException: org.springframework.boot.context.embedded.AnnotationConfigEmbeddedWebApplicationContext
- field (class "com.vaadin.spring.server.SpringUIProvider", name: "webApplicationContext", type: "interface org.springframework.web.context.WebApplicationContext")
- object (class "com.vaadin.spring.server.SpringUIProvider", com.vaadin.spring.server.SpringUIProvider@d30bfca)
- custom writeObject data (class "java.util.LinkedList")
- object (class "java.util.LinkedList", [com.vaadin.spring.server.SpringUIProvider@d30bfca])
- field (class "com.vaadin.server.VaadinSession", name: "uiProviders", type: "class java.util.LinkedList")
- root object (class "com.vaadin.server.VaadinSession", com.vaadin.server.VaadinSession@6241ff1a)
at org.springframework.data.redis.serializer.JdkSerializationRedisSerializer.serialize(JdkSerializationRedisSerializer.java:52)
at org.springframework.data.redis.core.AbstractOperations.rawHashValue(AbstractOperations.java:146)
at org.springframework.data.redis.core.DefaultHashOperations.putAll(DefaultHashOperations.java:128)
at org.springframework.data.redis.core.DefaultBoundHashOperations.putAll(DefaultBoundHashOperations.java:85)
at ...
```
Discussion from Skype:
[13/01/16 08:48:42] Johannes Tuikkala: It seem like that this is the culbrit for the serialization issue in hte SpringUIProvider private final WebApplicationContext webApplicationContext;
[13/01/16 08:50:19] Leif Åstrand: The solution in that case is most likely to make the field transient and non-final and add a custom private void readObject(java.io.ObjectInputStream stream) method that restores the WebApplicationContext reference from somewhere
[13/01/16 08:52:44] Leif Åstrand: There might also be other classes that hold references to WebApplicationContext or some other non-serializable spring class. Those will show up one after the other as you fix the previous one :)
Related ticket: #18164
Related StackOverflow thread: http://stackoverflow.com/questions/29388542/notserializableexception-on-serialization-of-objects-currently-shown-by-vaadin
I will try to make a test application for this, where this is easily reproducible.
___
Migrated from https://dev.vaadin.com/ticket/19462 | 1.0 | SpringUIProvider's webApplicationContext field is not serializable - **Originally by _johannes.tuikkala_ on 2016-01-13 09:02:40 +0200**
___
```
org.springframework.data.redis.serializer.SerializationException: Cannot serialize; nested exception is org.springframework.core.serializer.support.SerializationFailedException: Failed to serialize object using DefaultSerializer; nested exception is java.io.NotSerializableException: org.springframework.boot.context.embedded.AnnotationConfigEmbeddedWebApplicationContext
- field (class "com.vaadin.spring.server.SpringUIProvider", name: "webApplicationContext", type: "interface org.springframework.web.context.WebApplicationContext")
- object (class "com.vaadin.spring.server.SpringUIProvider", com.vaadin.spring.server.SpringUIProvider@d30bfca)
- custom writeObject data (class "java.util.LinkedList")
- object (class "java.util.LinkedList", [com.vaadin.spring.server.SpringUIProvider@d30bfca])
- field (class "com.vaadin.server.VaadinSession", name: "uiProviders", type: "class java.util.LinkedList")
- root object (class "com.vaadin.server.VaadinSession", com.vaadin.server.VaadinSession@6241ff1a)
at org.springframework.data.redis.serializer.JdkSerializationRedisSerializer.serialize(JdkSerializationRedisSerializer.java:52)
at org.springframework.data.redis.core.AbstractOperations.rawHashValue(AbstractOperations.java:146)
at org.springframework.data.redis.core.DefaultHashOperations.putAll(DefaultHashOperations.java:128)
at org.springframework.data.redis.core.DefaultBoundHashOperations.putAll(DefaultBoundHashOperations.java:85)
at ...
```
Discussion from Skype:
[13/01/16 08:48:42] Johannes Tuikkala: It seem like that this is the culbrit for the serialization issue in hte SpringUIProvider private final WebApplicationContext webApplicationContext;
[13/01/16 08:50:19] Leif Åstrand: The solution in that case is most likely to make the field transient and non-final and add a custom private void readObject(java.io.ObjectInputStream stream) method that restores the WebApplicationContext reference from somewhere
[13/01/16 08:52:44] Leif Åstrand: There might also be other classes that hold references to WebApplicationContext or some other non-serializable spring class. Those will show up one after the other as you fix the previous one :)
Related ticket: #18164
Related StackOverflow thread: http://stackoverflow.com/questions/29388542/notserializableexception-on-serialization-of-objects-currently-shown-by-vaadin
I will try to make a test application for this, where this is easily reproducible.
___
Migrated from https://dev.vaadin.com/ticket/19462 | defect | springuiprovider s webapplicationcontext field is not serializable originally by johannes tuikkala on org springframework data redis serializer serializationexception cannot serialize nested exception is org springframework core serializer support serializationfailedexception failed to serialize object using defaultserializer nested exception is java io notserializableexception org springframework boot context embedded annotationconfigembeddedwebapplicationcontext field class com vaadin spring server springuiprovider name webapplicationcontext type interface org springframework web context webapplicationcontext object class com vaadin spring server springuiprovider com vaadin spring server springuiprovider custom writeobject data class java util linkedlist object class java util linkedlist field class com vaadin server vaadinsession name uiproviders type class java util linkedlist root object class com vaadin server vaadinsession com vaadin server vaadinsession at org springframework data redis serializer jdkserializationredisserializer serialize jdkserializationredisserializer java at org springframework data redis core abstractoperations rawhashvalue abstractoperations java at org springframework data redis core defaulthashoperations putall defaulthashoperations java at org springframework data redis core defaultboundhashoperations putall defaultboundhashoperations java at discussion from skype johannes tuikkala it seem like that this is the culbrit for the serialization issue in hte springuiprovider private final webapplicationcontext webapplicationcontext leif åstrand the solution in that case is most likely to make the field transient and non final and add a custom private void readobject java io objectinputstream stream method that restores the webapplicationcontext reference from somewhere leif åstrand there might also be other classes that hold references to webapplicationcontext or some other non serializable spring class those will show up one after the other as you fix the previous one related ticket related stackoverflow thread i will try to make a test application for this where this is easily reproducible migrated from | 1 |
18,589 | 3,073,885,429 | IssuesEvent | 2015-08-20 01:27:28 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | opened | VM service: unclear how to invoke removeBreakpoint | Area-Observatory Type-Defect | The VM service documentation includes [a `removeBreakpoint()` function](https://github.com/dart-lang/sdk/blob/master/runtime/vm/service/service.md#removebreakpoint) that takes an isolate ID and a (string) breakpoint ID. However, [the `Breakpoint` class itself](https://github.com/dart-lang/sdk/blob/master/runtime/vm/service/service.md#breakpoint) doesn't include an ID field. The closest thing is the breakpoint number, but that's an integer, not a string. | 1.0 | VM service: unclear how to invoke removeBreakpoint - The VM service documentation includes [a `removeBreakpoint()` function](https://github.com/dart-lang/sdk/blob/master/runtime/vm/service/service.md#removebreakpoint) that takes an isolate ID and a (string) breakpoint ID. However, [the `Breakpoint` class itself](https://github.com/dart-lang/sdk/blob/master/runtime/vm/service/service.md#breakpoint) doesn't include an ID field. The closest thing is the breakpoint number, but that's an integer, not a string. | defect | vm service unclear how to invoke removebreakpoint the vm service documentation includes that takes an isolate id and a string breakpoint id however doesn t include an id field the closest thing is the breakpoint number but that s an integer not a string | 1 |
288,096 | 24,882,768,284 | IssuesEvent | 2022-10-28 03:47:08 | MPMG-DCC-UFMG/F01 | https://api.github.com/repos/MPMG-DCC-UFMG/F01 | closed | Teste de generalizacao para a tag Orçamento - Execução - Araçaí | generalization test development template - Memory (66) tag - Orçamento subtag - Execução | DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Araçaí. | 1.0 | Teste de generalizacao para a tag Orçamento - Execução - Araçaí - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Araçaí. | non_defect | teste de generalizacao para a tag orçamento execução araçaí dod realizar o teste de generalização do validador da tag orçamento execução para o município de araçaí | 0 |
417,904 | 28,111,625,423 | IssuesEvent | 2023-03-31 07:40:22 | quarto-dev/quarto-cli | https://api.github.com/repos/quarto-dev/quarto-cli | closed | Quarto CLI documentation in quarto.org | documentation epic | ### Discussed in https://github.com/quarto-dev/quarto-cli/discussions/4967
<div type='discussions-op-text'>
<sup>Originally posted by **mcanouil** March 25, 2023</sup>
The website is great regarding markdown typesetting.
In addition to what is already in it, I think it would be really great to also include Quarto CLI commands documentation as some kind of reference page (and/or "man" page).
Also, I noticed the `quarto use` command is not at all documented, not even in Quarto CLI itself.
<img width="1124" alt="image" src="https://user-images.githubusercontent.com/8896044/227738301-58da899b-4ecb-49fa-95f8-25adaa907bcd.png">
<img width="1124" alt="image" src="https://user-images.githubusercontent.com/8896044/227738347-0e25d47e-44a3-422c-a918-fc2fc2736978.png">
</div> | 1.0 | Quarto CLI documentation in quarto.org - ### Discussed in https://github.com/quarto-dev/quarto-cli/discussions/4967
<div type='discussions-op-text'>
<sup>Originally posted by **mcanouil** March 25, 2023</sup>
The website is great regarding markdown typesetting.
In addition to what is already in it, I think it would be really great to also include Quarto CLI commands documentation as some kind of reference page (and/or "man" page).
Also, I noticed the `quarto use` command is not at all documented, not even in Quarto CLI itself.
<img width="1124" alt="image" src="https://user-images.githubusercontent.com/8896044/227738301-58da899b-4ecb-49fa-95f8-25adaa907bcd.png">
<img width="1124" alt="image" src="https://user-images.githubusercontent.com/8896044/227738347-0e25d47e-44a3-422c-a918-fc2fc2736978.png">
</div> | non_defect | quarto cli documentation in quarto org discussed in originally posted by mcanouil march the website is great regarding markdown typesetting in addition to what is already in it i think it would be really great to also include quarto cli commands documentation as some kind of reference page and or man page also i noticed the quarto use command is not at all documented not even in quarto cli itself img width alt image src img width alt image src | 0 |
38,421 | 2,847,061,476 | IssuesEvent | 2015-05-29 15:01:41 | e-government-ua/i | https://api.github.com/repos/e-government-ua/i | opened | Создать сущность HistoryEvent (журнал событий) + REST-сервисы, по образцу реализованной сущности Document | hi priority | ВАЖНО: Использовать исключительно Hibernate+Liquibase+DAO+REST+JSON, как это реализовано уже для сущности Document (и автосоздание структуры с заливкой данных):
\i\wf-central\src\main\resources\liquibase\init.xml - авто-инициализация структуры через liquibase
\i\wf-central\src\main\resources\data\Document.csv - авто-заполнение структуры через liquibase
\i\wf-central\src\main\java\org\wf\dp\dniprorada\model\Document.java - сущность Hibernate
\i\wf-central\src\main\java\org\wf\dp\dniprorada\dao\DocumentDao.java - сущность DAO (ядро)
\i\wf-central\src\main\java\org\wf\dp\dniprorada\dao\DocumentDaoImpl.java - сущность DAO
\i\wf-central\src\main\java\org\activiti\rest\controller\ActivitiRestDocumentController.java - REST (/getDocument")
1) Создать сущность(таблицу) HistoryEvent, с такими полями:
nID (лонг) - номер-ИД (уникальный) //обязательное (автоитеррируемый)
nID_Subject (лонг) - номер-ИД субьекта, по которому событие //не обязательное (по умолчанию 0)
nID_HistoryEventType (лонг) - номер-ИД типа события, по которому событие //не обязательное (по умолчанию 0)
sEventName_Custom (<200 символов) - строка названия события (если номер-ИД типа события = 0), //не обязательное (по умолчанию пустая строка)
sMessage (безразмерный тип text) - строка сообщения //обязательное
sDate (тип date с милисекундами) - дата события //обязательное (автопроставляемая при добавлении записи)
2) Создать enum HistoryEventType, с такими полями:
nID - номер-ИД (уникальный)
sID - строка-ИД (уникальный)
sName - Название (на украинском)
sTemplate - Название (на украинском)
и заполнить, такими элементами:
nID=1, sID=getService
sName=Пользователь воспользовался услугой на портале
sTemplate=Ви подали заявку на послугу %Назва послуги%.
Статус: %статус%
nID=2, sID=setDocument_internal
sName=В Мои документы пользователя загружен новый документ – через наш портал
sTemplate=%Назва органу% завантажує %Назва документу% у Ваш розділ Мої документи
nID=3, sID=setDocument_external
sName=В Мои документы пользователя загружен новый документ – внешняя организация
sTemplate=%Назва органу% завантажує %Назва документу% у Ваш розділ Мої документи
nID=4, sID=setDocumentAccessLink
sName=Пользователь предоставил доступ к своему документу
sTemplate=Ви надаєте доступ до документу %Назва документу% іншій людині: %Ім’я того, кому надають доступ% (телефон: %телефон%)
nID=5, sID=setDocumentAccess
sName=Кто-то воспользовался доступом к документу через OTP, который ему предоставил пользователь
sTemplate=%Ім’я того, кому надають доступ% скористався доступом, який Ви надали, та переглянув документ %Назва документу%
3) Создать в контроллере ActivitiRestDocumentController.java новые REST-методы:
3.1. /getHistoryEvents (nID_Subject) - ответ массив объектов oHistoryEvent
3.2. /getHistoryEvent (nID) - ответ объект oHistoryEvent
3.3. /setHistoryEvent (все поля сущности, кроме nID) - ответ {"nID":54235} (c ИД-шником созданного обьекта oHistoryEvent)
и соблюсти в п.3. образец объекта(обязательно сохранить кейс и точные названия полей):
oHistoryEvent:{
"nID": 445
,"nID_Subject": 2
,"nID_HistoryEventType":3
,"sEventName":"" - консолидированное название события
,"sMessage":""
,"sDate":"2015-11-25 23:23:59.325"
}
5) Привязать поле nID_HistoryEventType сущности HistoryEvent к nID enum-а HistoryEventType, и по ней отображать в поле "sEventName" (обьекта oHistoryEvent) - название из поля sName (enum-а HistoryEventType), или, если nID_HistoryEventType=0, то значение из поля sEventName_Custom (своей-же сущности HistoryEvent)
6) Наполнить \i\wf-central\src\main\resources\data\HistoryEvent.csv десятком примеров (и обязательно подвязать все в liquibase\init.xml
| 1.0 | Создать сущность HistoryEvent (журнал событий) + REST-сервисы, по образцу реализованной сущности Document - ВАЖНО: Использовать исключительно Hibernate+Liquibase+DAO+REST+JSON, как это реализовано уже для сущности Document (и автосоздание структуры с заливкой данных):
\i\wf-central\src\main\resources\liquibase\init.xml - авто-инициализация структуры через liquibase
\i\wf-central\src\main\resources\data\Document.csv - авто-заполнение структуры через liquibase
\i\wf-central\src\main\java\org\wf\dp\dniprorada\model\Document.java - сущность Hibernate
\i\wf-central\src\main\java\org\wf\dp\dniprorada\dao\DocumentDao.java - сущность DAO (ядро)
\i\wf-central\src\main\java\org\wf\dp\dniprorada\dao\DocumentDaoImpl.java - сущность DAO
\i\wf-central\src\main\java\org\activiti\rest\controller\ActivitiRestDocumentController.java - REST (/getDocument")
1) Создать сущность(таблицу) HistoryEvent, с такими полями:
nID (лонг) - номер-ИД (уникальный) //обязательное (автоитеррируемый)
nID_Subject (лонг) - номер-ИД субьекта, по которому событие //не обязательное (по умолчанию 0)
nID_HistoryEventType (лонг) - номер-ИД типа события, по которому событие //не обязательное (по умолчанию 0)
sEventName_Custom (<200 символов) - строка названия события (если номер-ИД типа события = 0), //не обязательное (по умолчанию пустая строка)
sMessage (безразмерный тип text) - строка сообщения //обязательное
sDate (тип date с милисекундами) - дата события //обязательное (автопроставляемая при добавлении записи)
2) Создать enum HistoryEventType, с такими полями:
nID - номер-ИД (уникальный)
sID - строка-ИД (уникальный)
sName - Название (на украинском)
sTemplate - Название (на украинском)
и заполнить, такими элементами:
nID=1, sID=getService
sName=Пользователь воспользовался услугой на портале
sTemplate=Ви подали заявку на послугу %Назва послуги%.
Статус: %статус%
nID=2, sID=setDocument_internal
sName=В Мои документы пользователя загружен новый документ – через наш портал
sTemplate=%Назва органу% завантажує %Назва документу% у Ваш розділ Мої документи
nID=3, sID=setDocument_external
sName=В Мои документы пользователя загружен новый документ – внешняя организация
sTemplate=%Назва органу% завантажує %Назва документу% у Ваш розділ Мої документи
nID=4, sID=setDocumentAccessLink
sName=Пользователь предоставил доступ к своему документу
sTemplate=Ви надаєте доступ до документу %Назва документу% іншій людині: %Ім’я того, кому надають доступ% (телефон: %телефон%)
nID=5, sID=setDocumentAccess
sName=Кто-то воспользовался доступом к документу через OTP, который ему предоставил пользователь
sTemplate=%Ім’я того, кому надають доступ% скористався доступом, який Ви надали, та переглянув документ %Назва документу%
3) Создать в контроллере ActivitiRestDocumentController.java новые REST-методы:
3.1. /getHistoryEvents (nID_Subject) - ответ массив объектов oHistoryEvent
3.2. /getHistoryEvent (nID) - ответ объект oHistoryEvent
3.3. /setHistoryEvent (все поля сущности, кроме nID) - ответ {"nID":54235} (c ИД-шником созданного обьекта oHistoryEvent)
и соблюсти в п.3. образец объекта(обязательно сохранить кейс и точные названия полей):
oHistoryEvent:{
"nID": 445
,"nID_Subject": 2
,"nID_HistoryEventType":3
,"sEventName":"" - консолидированное название события
,"sMessage":""
,"sDate":"2015-11-25 23:23:59.325"
}
5) Привязать поле nID_HistoryEventType сущности HistoryEvent к nID enum-а HistoryEventType, и по ней отображать в поле "sEventName" (обьекта oHistoryEvent) - название из поля sName (enum-а HistoryEventType), или, если nID_HistoryEventType=0, то значение из поля sEventName_Custom (своей-же сущности HistoryEvent)
6) Наполнить \i\wf-central\src\main\resources\data\HistoryEvent.csv десятком примеров (и обязательно подвязать все в liquibase\init.xml
| non_defect | создать сущность historyevent журнал событий rest сервисы по образцу реализованной сущности document важно использовать исключительно hibernate liquibase dao rest json как это реализовано уже для сущности document и автосоздание структуры с заливкой данных i wf central src main resources liquibase init xml авто инициализация структуры через liquibase i wf central src main resources data document csv авто заполнение структуры через liquibase i wf central src main java org wf dp dniprorada model document java сущность hibernate i wf central src main java org wf dp dniprorada dao documentdao java сущность dao ядро i wf central src main java org wf dp dniprorada dao documentdaoimpl java сущность dao i wf central src main java org activiti rest controller activitirestdocumentcontroller java rest getdocument создать сущность таблицу historyevent с такими полями nid лонг номер ид уникальный обязательное автоитеррируемый nid subject лонг номер ид субьекта по которому событие не обязательное по умолчанию nid historyeventtype лонг номер ид типа события по которому событие не обязательное по умолчанию seventname custom символов строка названия события если номер ид типа события не обязательное по умолчанию пустая строка smessage безразмерный тип text строка сообщения обязательное sdate тип date с милисекундами дата события обязательное автопроставляемая при добавлении записи создать enum historyeventtype с такими полями nid номер ид уникальный sid строка ид уникальный sname название на украинском stemplate название на украинском и заполнить такими элементами nid sid getservice sname пользователь воспользовался услугой на портале stemplate ви подали заявку на послугу назва послуги статус статус nid sid setdocument internal sname в мои документы пользователя загружен новый документ – через наш портал stemplate назва органу завантажує назва документу у ваш розділ мої документи nid sid setdocument external sname в мои документы пользователя загружен новый документ – внешняя организация stemplate назва органу завантажує назва документу у ваш розділ мої документи nid sid setdocumentaccesslink sname пользователь предоставил доступ к своему документу stemplate ви надаєте доступ до документу назва документу іншій людині ім’я того кому надають доступ телефон телефон nid sid setdocumentaccess sname кто то воспользовался доступом к документу через otp который ему предоставил пользователь stemplate ім’я того кому надають доступ скористався доступом який ви надали та переглянув документ назва документу создать в контроллере activitirestdocumentcontroller java новые rest методы gethistoryevents nid subject ответ массив объектов ohistoryevent gethistoryevent nid ответ объект ohistoryevent sethistoryevent все поля сущности кроме nid ответ nid c ид шником созданного обьекта ohistoryevent и соблюсти в п образец объекта обязательно сохранить кейс и точные названия полей ohistoryevent nid nid subject nid historyeventtype seventname консолидированное название события smessage sdate привязать поле nid historyeventtype сущности historyevent к nid enum а historyeventtype и по ней отображать в поле seventname обьекта ohistoryevent название из поля sname enum а historyeventtype или если nid historyeventtype то значение из поля seventname custom своей же сущности historyevent наполнить i wf central src main resources data historyevent csv десятком примеров и обязательно подвязать все в liquibase init xml | 0 |
73,424 | 24,620,579,411 | IssuesEvent | 2022-10-15 21:50:03 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Joining loop with group video Element Call | T-Defect | ### Steps to reproduce
Hello, I am in a joining loop. Whenever I press join in a lobby it kicks me back to the native remove (widget is gone). A card says "started a video call" and when I press "Join" the lobby reappears, then "join" breaks the widget again.
Encryption is turned within the room settings.
### Outcome
#### expected
1. Press video icon in banner
2. Choose "Element Call" from dropdown
3. Brought to lobby
4. Press "Join"
5. entered into group call
### actual
1. Press video icon in banner
2. Choose "Element Call" from dropdown
3. Brought to lobby
4. Press "Join" and nothing visually changes
6. Widget is removed from screen and room appears
7. Cannot enter call
### Operating system
Debian 11
### Application version
Element Nightly version: 2022101501 Olm version: 3.2.12
### How did you install the app?
element deb repo
### Homeserver
ems.one.host
### Will you send logs?
Yes | 1.0 | Joining loop with group video Element Call - ### Steps to reproduce
Hello, I am in a joining loop. Whenever I press join in a lobby it kicks me back to the native remove (widget is gone). A card says "started a video call" and when I press "Join" the lobby reappears, then "join" breaks the widget again.
Encryption is turned within the room settings.
### Outcome
#### expected
1. Press video icon in banner
2. Choose "Element Call" from dropdown
3. Brought to lobby
4. Press "Join"
5. entered into group call
### actual
1. Press video icon in banner
2. Choose "Element Call" from dropdown
3. Brought to lobby
4. Press "Join" and nothing visually changes
6. Widget is removed from screen and room appears
7. Cannot enter call
### Operating system
Debian 11
### Application version
Element Nightly version: 2022101501 Olm version: 3.2.12
### How did you install the app?
element deb repo
### Homeserver
ems.one.host
### Will you send logs?
Yes | defect | joining loop with group video element call steps to reproduce hello i am in a joining loop whenever i press join in a lobby it kicks me back to the native remove widget is gone a card says started a video call and when i press join the lobby reappears then join breaks the widget again encryption is turned within the room settings outcome expected press video icon in banner choose element call from dropdown brought to lobby press join entered into group call actual press video icon in banner choose element call from dropdown brought to lobby press join and nothing visually changes widget is removed from screen and room appears cannot enter call operating system debian application version element nightly version olm version how did you install the app element deb repo homeserver ems one host will you send logs yes | 1 |
31,316 | 6,497,778,385 | IssuesEvent | 2017-08-22 15:04:08 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | [discovery] HazelcastClient always tries to connect to localhost when using DiscoveryStrategy | Team: Client Type: Defect | (HazelcastClient version 3.8)
When using the Discovery system to locate nodes the `DefaultAddressProvider` still get added to the addressProviders list causing the hazelcast(client) to try to connect to localhost. This is both unneeded and unwanted when using for example the `MulticastDiscoveryStrategy`.
In `com.hazelcast.client.impl.HazelcastClientInstanceImpl.createAddressProviders ~ line 265`
the `DefaultAddressProvider` get added which has a instance of `ClientNetworkConfig`
Because `ClientNetworkConfig ~ line 212` alway returns a list with localhost if nothing is specified we end up with a Client that always connect to localhost unless a explict address is provided (which is unwanted in many discovery cases).
Proposed solution:
only include `DefaultAddresProviders` if no others are provided.
psuedo patch:
```java
src/main/java/com/hazelcast/client/impl/HazelcastClientInstanceImpl.java
- 264: addressProviders.add(new DefaultAddressProvider(networkConfig));
+ 281: if (addressProviders.isEmpty()) {
+ 282 addressProviders.add(new DefaultAddressProvider(networkConfig));
+ 283: }
```
Reproducable by using any DiscoveryStrategy and not adding addresses to the clientconfig.xml
```xml
example client config:
<hazelcast-client xsi:schemaLocation="http://www.hazelcast.com/schema/config hazelcast-config-3.8.xsd"
xmlns="http://www.hazelcast.com/schema/client-config"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<group>
<name>hazelcast-dev</name>
<password>dev-pass</password>
</group>
<network>
<aws enabled="false"/>
<connection-timeout>10000</connection-timeout>
<connection-attempt-period>500</connection-attempt-period>
<connection-attempt-limit>20000</connection-attempt-limit>
<smart-routing>true</smart-routing>
<redo-operation>true</redo-operation>
<discovery-strategies>
<discovery-strategy class="com.hazelcast.spi.discovery.multicast.MulticastDiscoveryStrategy" enabled="true" >
<properties>
<property name="port" >9999</property>
<property name="group" >230.18.0.255</property>
</properties>
</discovery-strategy>
</discovery-strategies>
</network>
``` | 1.0 | [discovery] HazelcastClient always tries to connect to localhost when using DiscoveryStrategy - (HazelcastClient version 3.8)
When using the Discovery system to locate nodes the `DefaultAddressProvider` still get added to the addressProviders list causing the hazelcast(client) to try to connect to localhost. This is both unneeded and unwanted when using for example the `MulticastDiscoveryStrategy`.
In `com.hazelcast.client.impl.HazelcastClientInstanceImpl.createAddressProviders ~ line 265`
the `DefaultAddressProvider` get added which has a instance of `ClientNetworkConfig`
Because `ClientNetworkConfig ~ line 212` alway returns a list with localhost if nothing is specified we end up with a Client that always connect to localhost unless a explict address is provided (which is unwanted in many discovery cases).
Proposed solution:
only include `DefaultAddresProviders` if no others are provided.
psuedo patch:
```java
src/main/java/com/hazelcast/client/impl/HazelcastClientInstanceImpl.java
- 264: addressProviders.add(new DefaultAddressProvider(networkConfig));
+ 281: if (addressProviders.isEmpty()) {
+ 282 addressProviders.add(new DefaultAddressProvider(networkConfig));
+ 283: }
```
Reproducable by using any DiscoveryStrategy and not adding addresses to the clientconfig.xml
```xml
example client config:
<hazelcast-client xsi:schemaLocation="http://www.hazelcast.com/schema/config hazelcast-config-3.8.xsd"
xmlns="http://www.hazelcast.com/schema/client-config"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<group>
<name>hazelcast-dev</name>
<password>dev-pass</password>
</group>
<network>
<aws enabled="false"/>
<connection-timeout>10000</connection-timeout>
<connection-attempt-period>500</connection-attempt-period>
<connection-attempt-limit>20000</connection-attempt-limit>
<smart-routing>true</smart-routing>
<redo-operation>true</redo-operation>
<discovery-strategies>
<discovery-strategy class="com.hazelcast.spi.discovery.multicast.MulticastDiscoveryStrategy" enabled="true" >
<properties>
<property name="port" >9999</property>
<property name="group" >230.18.0.255</property>
</properties>
</discovery-strategy>
</discovery-strategies>
</network>
``` | defect | hazelcastclient always tries to connect to localhost when using discoverystrategy hazelcastclient version when using the discovery system to locate nodes the defaultaddressprovider still get added to the addressproviders list causing the hazelcast client to try to connect to localhost this is both unneeded and unwanted when using for example the multicastdiscoverystrategy in com hazelcast client impl hazelcastclientinstanceimpl createaddressproviders line the defaultaddressprovider get added which has a instance of clientnetworkconfig because clientnetworkconfig line alway returns a list with localhost if nothing is specified we end up with a client that always connect to localhost unless a explict address is provided which is unwanted in many discovery cases proposed solution only include defaultaddresproviders if no others are provided psuedo patch java src main java com hazelcast client impl hazelcastclientinstanceimpl java addressproviders add new defaultaddressprovider networkconfig if addressproviders isempty addressproviders add new defaultaddressprovider networkconfig reproducable by using any discoverystrategy and not adding addresses to the clientconfig xml xml example client config hazelcast client xsi schemalocation hazelcast config xsd xmlns xmlns xsi hazelcast dev dev pass true true | 1 |
417,743 | 28,110,902,924 | IssuesEvent | 2023-03-31 07:05:27 | daniellingzy/ped | https://api.github.com/repos/daniellingzy/ped | opened | UG Bug- Summary statistic box component differs from the screenshots in the UG | type.DocumentationBug severity.Medium | Actual display upon following quick start instructions:

Expected display based on quickstart instructions:

Note how the summary statistic box has differing display and components. Please do standardise upon one variation.
<!--session: 1680243625937-003781c7-eb21-4c47-b057-4340a5c3a047-->
<!--Version: Web v3.4.7--> | 1.0 | UG Bug- Summary statistic box component differs from the screenshots in the UG - Actual display upon following quick start instructions:

Expected display based on quickstart instructions:

Note how the summary statistic box has differing display and components. Please do standardise upon one variation.
<!--session: 1680243625937-003781c7-eb21-4c47-b057-4340a5c3a047-->
<!--Version: Web v3.4.7--> | non_defect | ug bug summary statistic box component differs from the screenshots in the ug actual display upon following quick start instructions expected display based on quickstart instructions note how the summary statistic box has differing display and components please do standardise upon one variation | 0 |
6,552 | 2,850,548,471 | IssuesEvent | 2015-05-31 17:30:51 | MeteorCode/Pathway | https://api.github.com/repos/MeteorCode/Pathway | closed | Silence log messages during testing | testing | Currently, the [console spam](https://travis-ci.org/MeteorCode/Pathway/builds/61902801#L266) from `Context` during testing is way obnoxious; it prints almost 2000 lines of log message. This makes reading the build log a pain. Need to make it stop. | 1.0 | Silence log messages during testing - Currently, the [console spam](https://travis-ci.org/MeteorCode/Pathway/builds/61902801#L266) from `Context` during testing is way obnoxious; it prints almost 2000 lines of log message. This makes reading the build log a pain. Need to make it stop. | non_defect | silence log messages during testing currently the from context during testing is way obnoxious it prints almost lines of log message this makes reading the build log a pain need to make it stop | 0 |
75,989 | 15,495,734,352 | IssuesEvent | 2021-03-11 01:23:47 | haeli05/source | https://api.github.com/repos/haeli05/source | opened | CVE-2019-16776 (High) detected in npm-6.11.3.tgz | security vulnerability | ## CVE-2019-16776 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-6.11.3.tgz</b></p></summary>
<p>a package manager for JavaScript</p>
<p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-6.11.3.tgz">https://registry.npmjs.org/npm/-/npm-6.11.3.tgz</a></p>
<p>Path to dependency file: source/FrontEnd/package.json</p>
<p>Path to vulnerable library: source/FrontEnd/node_modules/npm/package.json</p>
<p>
Dependency Hierarchy:
- :x: **npm-6.11.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.
<p>Publish Date: 2019-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p>
<p>Release Date: 2019-12-13</p>
<p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-16776 (High) detected in npm-6.11.3.tgz - ## CVE-2019-16776 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-6.11.3.tgz</b></p></summary>
<p>a package manager for JavaScript</p>
<p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-6.11.3.tgz">https://registry.npmjs.org/npm/-/npm-6.11.3.tgz</a></p>
<p>Path to dependency file: source/FrontEnd/package.json</p>
<p>Path to vulnerable library: source/FrontEnd/node_modules/npm/package.json</p>
<p>
Dependency Hierarchy:
- :x: **npm-6.11.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.
<p>Publish Date: 2019-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p>
<p>Release Date: 2019-12-13</p>
<p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in npm tgz cve high severity vulnerability vulnerable library npm tgz a package manager for javascript library home page a href path to dependency file source frontend package json path to vulnerable library source frontend node modules npm package json dependency hierarchy x npm tgz vulnerable library vulnerability details versions of the npm cli prior to are vulnerable to an arbitrary file write it fails to prevent access to folders outside of the intended node modules folder through the bin field a properly constructed entry in the package json bin field would allow a package publisher to modify and or gain access to arbitrary files on a user s system when the package is installed this behavior is still possible through install scripts this vulnerability bypasses a user using the ignore scripts install option publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution npm yarn step up your open source security game with whitesource | 0 |
139,273 | 31,390,394,543 | IssuesEvent | 2023-08-26 09:03:51 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Option "Browser Page Title" not working | Feature No Code Attached Yet | ### Steps to reproduce the issue
I create an article and add a title into the field "Browser Page Title" in the tab "Options" - e.g. "xxx my SEO optimized title-text xxx".
Then I create a menu-item for this article. Here (in the menu-item) I do NOT add text into the field "Browser Page Title" in the tab "Page Display", the field is still empty.
### Expected result
The title-tag (<title></title>) in the code of the frontend should show the title which I have set into the article, into the field "Browser Page Title" (= xxx my SEO optimized title-text xxx).
### Actual result
It shows the wront text in the title-tag (<title></title>): It shows the "MENU title" (= the wording of the textlink in the menu) instead of the "Browser Page Title" (= xxx my SEO optimized title-text xxx).
### System information (as much as possible)
Joomla! 4.3.3 with YOOtheme Version 4.0.10
### Additional comments
I refer to the issue [#28468 ](https://github.com/joomla/joomla-cms/issues/28468) . a comment here says that the issue is solved: https://github.com/joomla/joomla-cms/pull/39249. but it isn't... | 1.0 | Option "Browser Page Title" not working - ### Steps to reproduce the issue
I create an article and add a title into the field "Browser Page Title" in the tab "Options" - e.g. "xxx my SEO optimized title-text xxx".
Then I create a menu-item for this article. Here (in the menu-item) I do NOT add text into the field "Browser Page Title" in the tab "Page Display", the field is still empty.
### Expected result
The title-tag (<title></title>) in the code of the frontend should show the title which I have set into the article, into the field "Browser Page Title" (= xxx my SEO optimized title-text xxx).
### Actual result
It shows the wront text in the title-tag (<title></title>): It shows the "MENU title" (= the wording of the textlink in the menu) instead of the "Browser Page Title" (= xxx my SEO optimized title-text xxx).
### System information (as much as possible)
Joomla! 4.3.3 with YOOtheme Version 4.0.10
### Additional comments
I refer to the issue [#28468 ](https://github.com/joomla/joomla-cms/issues/28468) . a comment here says that the issue is solved: https://github.com/joomla/joomla-cms/pull/39249. but it isn't... | non_defect | option browser page title not working steps to reproduce the issue i create an article and add a title into the field browser page title in the tab options e g xxx my seo optimized title text xxx then i create a menu item for this article here in the menu item i do not add text into the field browser page title in the tab page display the field is still empty expected result the title tag in the code of the frontend should show the title which i have set into the article into the field browser page title xxx my seo optimized title text xxx actual result it shows the wront text in the title tag it shows the menu title the wording of the textlink in the menu instead of the browser page title xxx my seo optimized title text xxx system information as much as possible joomla with yootheme version additional comments i refer to the issue a comment here says that the issue is solved but it isn t | 0 |
45,870 | 13,055,812,157 | IssuesEvent | 2020-07-30 02:48:41 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | dataclasses std_map_indexing_suite_* tests failing on EL4/Python 2.3 (Trac #162) | Incomplete Migration Migrated from Trac dataclasses defect | Migrated from https://code.icecube.wisc.edu/ticket/162
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "these test scripts are failing on EL4/SL4 with python 2.3.4\n\nMaybe this python is just too old. But should check.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1416713876862109",
"component": "dataclasses",
"summary": "dataclasses std_map_indexing_suite_* tests failing on EL4/Python 2.3",
"priority": "normal",
"keywords": "",
"time": "2009-06-12T20:41:25",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
| 1.0 | dataclasses std_map_indexing_suite_* tests failing on EL4/Python 2.3 (Trac #162) - Migrated from https://code.icecube.wisc.edu/ticket/162
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "these test scripts are failing on EL4/SL4 with python 2.3.4\n\nMaybe this python is just too old. But should check.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1416713876862109",
"component": "dataclasses",
"summary": "dataclasses std_map_indexing_suite_* tests failing on EL4/Python 2.3",
"priority": "normal",
"keywords": "",
"time": "2009-06-12T20:41:25",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
| defect | dataclasses std map indexing suite tests failing on python trac migrated from json status closed changetime description these test scripts are failing on with python n nmaybe this python is just too old but should check reporter blaufuss cc resolution fixed ts component dataclasses summary dataclasses std map indexing suite tests failing on python priority normal keywords time milestone owner jvansanten type defect | 1 |
1,496 | 2,603,966,530 | IssuesEvent | 2015-02-24 18:59:11 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳辽宁治疗疱疹医院 | auto-migrated Priority-Medium Type-Defect | ```
沈阳辽宁治疗疱疹医院〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:04 | 1.0 | 沈阳辽宁治疗疱疹医院 - ```
沈阳辽宁治疗疱疹医院〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:04 | defect | 沈阳辽宁治疗疱疹医院 沈阳辽宁治疗疱疹医院〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at | 1 |
35,398 | 7,727,668,426 | IssuesEvent | 2018-05-25 04:10:07 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | SortedList implementation causes Javascript runtime error | defect in-progress | I copied dotnet source for sortedlist implementation, and removed the resource string. The resulting code compiles fine, but Javascript execution fails.
I tracked the issue down to the sub-class
private struct Enumerator : IEnumerator<KeyValuePair<TK, TV>>, IDictionaryEnumerator
If I remove the Enumerator class, runtime error goes away but of course the sortedlist does not work without the enumerator.
Further, If I leave the sub-class, and remove IDictionaryEnumerator, then it works.
### Steps To Reproduce
https://deck.net/94fd47356bddf4c009b256a472561736
```csharp
public class Program
{
public class SortedList<TK, TV> :
IDictionary<TK, TV>
{
private TK[] keys; // Do not rename (binary serialization)
private TV[] values; // Do not rename (binary serialization)
private int _size; // Do not rename (binary serialization)
private int version; // Do not rename (binary serialization)
private IComparer<TK> comparer; // Do not rename (binary serialization)
private KeyList keyList; // Do not rename (binary serialization)
private ValueList valueList; // Do not rename (binary serialization)
[NonSerialized]
private object _syncRoot;
private const int DefaultCapacity = 4;
// Constructs a new sorted list. The sorted list is initially empty and has
// a capacity of zero. Upon adding the first element to the sorted list the
// capacity is increased to DefaultCapacity, and then increased in multiples of two as
// required. The elements of the sorted list are ordered according to the
// IComparable interface, which must be implemented by the keys of
// all entries added to the sorted list.
public SortedList()
{
keys = new TK[] { };
values = new TV[] { };
_size = 0;
comparer = Comparer<TK>.Default;
}
// Constructs a new sorted list. The sorted list is initially empty and has
// a capacity of zero. Upon adding the first element to the sorted list the
// capacity is increased to 16, and then increased in multiples of two as
// required. The elements of the sorted list are ordered according to the
// IComparable interface, which must be implemented by the keys of
// all entries added to the sorted list.
//
public SortedList(int capacity)
{
if (capacity < 0)
throw new ArgumentOutOfRangeException(nameof(capacity), capacity, "SR.ArgumentOutOfRange_NeedNonNegNum");
keys = new TK[capacity];
values = new TV[capacity];
comparer = Comparer<TK>.Default;
}
// Constructs a new sorted list with a given IComparer
// implementation. The sorted list is initially empty and has a capacity of
// zero. Upon adding the first element to the sorted list the capacity is
// increased to 16, and then increased in multiples of two as required. The
// elements of the sorted list are ordered according to the given
// IComparer implementation. If comparer is null, the
// elements are compared to each other using the IComparable
// interface, which in that case must be implemented by the keys of all
// entries added to the sorted list.
//
public SortedList(IComparer<TK> comparer)
: this()
{
if (comparer != null)
{
this.comparer = comparer;
}
}
// Constructs a new sorted dictionary with a given IComparer
// implementation and a given initial capacity. The sorted list is
// initially empty, but will have room for the given number of elements
// before any reallocations are required. The elements of the sorted list
// are ordered according to the given IComparer implementation. If
// comparer is null, the elements are compared to each other using
// the IComparable interface, which in that case must be implemented
// by the keys of all entries added to the sorted list.
//
public SortedList(int capacity, IComparer<TK> comparer)
: this(comparer)
{
Capacity = capacity;
}
// Constructs a new sorted list containing a copy of the entries in the
// given dictionary. The elements of the sorted list are ordered according
// to the IComparable interface, which must be implemented by the
// keys of all entries in the given dictionary as well as keys
// subsequently added to the sorted list.
//
public SortedList(IDictionary<TK, TV> dictionary)
: this(dictionary, null)
{
}
// Constructs a new sorted list containing a copy of the entries in the
// given dictionary. The elements of the sorted list are ordered according
// to the given IComparer implementation. If comparer is
// null, the elements are compared to each other using the
// IComparable interface, which in that case must be implemented
// by the keys of all entries in the given dictionary as well as keys
// subsequently added to the sorted list.
//
public SortedList(IDictionary<TK, TV> dictionary, IComparer<TK> comparer)
: this((dictionary != null ? dictionary.Count : 0), comparer)
{
if (dictionary == null)
throw new ArgumentNullException(nameof(dictionary));
int count = dictionary.Count;
if (count != 0)
{
TK[] keys = this.keys;
dictionary.Keys.CopyTo(keys, 0);
dictionary.Values.CopyTo(values, 0);
Debug.Assert(count == this.keys.Length);
if (count > 1)
{
comparer = Comparer; // obtain default if this is null.
keys = dictionary.OrderBy(x => x.Key, comparer).Select(x => x.Key).ToArray();
for (int i = 1; i != keys.Length; ++i)
{
if (comparer.Compare(keys[i - 1], keys[i]) == 0)
{
throw new ArgumentException("SR.Format(SR.Argument_AddingDuplicate, keys[i])");
}
}
}
}
_size = count;
}
// Adds an entry with the given key and value to this sorted list. An
// ArgumentException is thrown if the key is already present in the sorted list.
//
public void Add(TK key, TV value)
{
if (key == null) throw new ArgumentNullException(nameof(key));
int i = Array.BinarySearch<TK>(keys, 0, _size, key, comparer);
if (i >= 0)
throw new ArgumentException("SR.Format(SR.Argument_AddingDuplicate, key), nameof(key)");
Insert(~i, key, value);
}
void ICollection<KeyValuePair<TK, TV>>.Add(KeyValuePair<TK, TV> keyValuePair)
{
Add(keyValuePair.Key, keyValuePair.Value);
}
bool ICollection<KeyValuePair<TK, TV>>.Contains(KeyValuePair<TK, TV> keyValuePair)
{
int index = IndexOfKey(keyValuePair.Key);
if (index >= 0 && EqualityComparer<TV>.Default.Equals(values[index], keyValuePair.Value))
{
return true;
}
return false;
}
bool ICollection<KeyValuePair<TK, TV>>.Remove(KeyValuePair<TK, TV> keyValuePair)
{
int index = IndexOfKey(keyValuePair.Key);
if (index >= 0 && EqualityComparer<TV>.Default.Equals(values[index], keyValuePair.Value))
{
RemoveAt(index);
return true;
}
return false;
}
// Returns the capacity of this sorted list. The capacity of a sorted list
// represents the allocated length of the internal arrays used to store the
// keys and values of the list, and thus also indicates the maximum number
// of entries the list can contain before a reallocation of the internal
// arrays is required.
//
public int Capacity
{
get
{
return keys.Length;
}
set
{
if (value != keys.Length)
{
if (value < _size)
{
throw new ArgumentOutOfRangeException(nameof(value), value, "SR.ArgumentOutOfRange_SmallCapacity");
}
if (value > 0)
{
TK[] newKeys = new TK[value];
TV[] newValues = new TV[value];
if (_size > 0)
{
Array.Copy(keys, 0, newKeys, 0, _size);
Array.Copy(values, 0, newValues, 0, _size);
}
keys = newKeys;
values = newValues;
}
else
{
keys = new TK[] { };
values = new TV[] { };
}
}
}
}
public IComparer<TK> Comparer
{
get
{
return comparer;
}
}
// Returns the number of entries in this sorted list.
public int Count
{
get
{
return _size;
}
}
// Returns a collection representing the keys of this sorted list. This
// method returns the same object as GetKeyList, but typed as an
// ICollection instead of an IList.
public IList<TK> Keys
{
get
{
return GetKeyListHelper();
}
}
ICollection<TK> IDictionary<TK, TV>.Keys
{
get
{
return GetKeyListHelper();
}
}
// Returns a collection representing the values of this sorted list. This
// method returns the same object as GetValueList, but typed as an
// ICollection instead of an IList.
//
public IList<TV> Values
{
get
{
return GetValueListHelper();
}
}
ICollection<TV> IDictionary<TK, TV>.Values
{
get
{
return GetValueListHelper();
}
}
private KeyList GetKeyListHelper()
{
if (keyList == null)
keyList = new KeyList(this);
return keyList;
}
private ValueList GetValueListHelper()
{
if (valueList == null)
valueList = new ValueList(this);
return valueList;
}
bool ICollection<KeyValuePair<TK, TV>>.IsReadOnly
{
get { return false; }
}
// Removes all entries from this sorted list.
public void Clear()
{
// clear does not change the capacity
version++;
// TODO:
// Don't need to doc this but we clear the elements so that the gc can reclaim the references.
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TKey>())
//{
// Array.Clear(keys, 0, _size);
//}
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TValue>())
//{
// Array.Clear(values, 0, _size);
//}
_size = 0;
}
// Checks if this sorted list contains an entry with the given key.
public bool ContainsKey(TK key)
{
return IndexOfKey(key) >= 0;
}
// Checks if this sorted list contains an entry with the given value. The
// values of the entries of the sorted list are compared to the given value
// using the Object.Equals method. This method performs a linear
// search and is substantially slower than the Contains
// method.
public bool ContainsValue(TV value)
{
return IndexOfValue(value) >= 0;
}
// Copies the values in this SortedList to an array.
void ICollection<KeyValuePair<TK, TV>>.CopyTo(KeyValuePair<TK, TV>[] array, int arrayIndex)
{
if (array == null)
{
throw new ArgumentNullException(nameof(array));
}
if (arrayIndex < 0 || arrayIndex > array.Length)
{
throw new ArgumentOutOfRangeException(nameof(arrayIndex), arrayIndex, "SR.ArgumentOutOfRange_Index");
}
if (array.Length - arrayIndex < Count)
{
throw new ArgumentException("SR.Arg_ArrayPlusOffTooSmall");
}
for (int i = 0; i < Count; i++)
{
KeyValuePair<TK, TV> entry = new KeyValuePair<TK, TV>(keys[i], values[i]);
array[arrayIndex + i] = entry;
}
}
private const int MaxArrayLength = 0X7FEFFFFF;
// Ensures that the capacity of this sorted list is at least the given
// minimum value. If the current capacity of the list is less than
// min, the capacity is increased to twice the current capacity or
// to min, whichever is larger.
private void EnsureCapacity(int min)
{
int newCapacity = keys.Length == 0 ? DefaultCapacity : keys.Length * 2;
// Allow the list to grow to maximum possible capacity (~2G elements) before encountering overflow.
// Note that this check works even when _items.Length overflowed thanks to the (uint) cast
if ((uint)newCapacity > MaxArrayLength) newCapacity = MaxArrayLength;
if (newCapacity < min) newCapacity = min;
Capacity = newCapacity;
}
// Returns the value of the entry at the given index.
private TV GetByIndex(int index)
{
if (index < 0 || index >= _size)
throw new ArgumentOutOfRangeException(nameof(index), index, "SR.ArgumentOutOfRange_Index");
return values[index];
}
public IEnumerator<KeyValuePair<TK, TV>> GetEnumerator()
{
return new Enumerator(this, Enumerator.KeyValuePair);
}
IEnumerator<KeyValuePair<TK, TV>> IEnumerable<KeyValuePair<TK, TV>>.GetEnumerator()
{
return new Enumerator(this, Enumerator.KeyValuePair);
}
IEnumerator IEnumerable.GetEnumerator()
{
return new Enumerator(this, Enumerator.KeyValuePair);
}
// Returns the key of the entry at the given index.
private TK GetKey(int index)
{
if (index < 0 || index >= _size)
throw new ArgumentOutOfRangeException(nameof(index), index, "SR.ArgumentOutOfRange_Index");
return keys[index];
}
// Returns the value associated with the given key. If an entry with the
// given key is not found, the returned value is null.
public TV this[TK key]
{
get
{
int i = IndexOfKey(key);
if (i >= 0)
return values[i];
throw new KeyNotFoundException("SR.Format(SR.Arg_KeyNotFoundWithKey, key.ToString())");
}
set
{
if (((object)key) == null) throw new ArgumentNullException(nameof(key));
int i = Array.BinarySearch<TK>(keys, 0, _size, key, comparer);
if (i >= 0)
{
values[i] = value;
version++;
return;
}
Insert(~i, key, value);
}
}
// Returns the index of the entry with a given key in this sorted list. The
// key is located through a binary search, and thus the average execution
// time of this method is proportional to Log2(size), where
// size is the size of this sorted list. The returned value is -1 if
// the given key does not occur in this sorted list. Null is an invalid
// key value.
public int IndexOfKey(TK key)
{
if (key == null)
throw new ArgumentNullException(nameof(key));
int ret = Array.BinarySearch<TK>(keys, 0, _size, key, comparer);
return ret >= 0 ? ret : -1;
}
// Returns the index of the first occurrence of an entry with a given value
// in this sorted list. The entry is located through a linear search, and
// thus the average execution time of this method is proportional to the
// size of this sorted list. The elements of the list are compared to the
// given value using the Object.Equals method.
public int IndexOfValue(TV value)
{
return Array.IndexOf(values, value, 0, _size);
}
// Inserts an entry with a given key and value at a given index.
private void Insert(int index, TK key, TV value)
{
if (_size == keys.Length) EnsureCapacity(_size + 1);
if (index < _size)
{
Array.Copy(keys, index, keys, index + 1, _size - index);
Array.Copy(values, index, values, index + 1, _size - index);
}
keys[index] = key;
values[index] = value;
_size++;
version++;
}
public bool TryGetValue(TK key, out TV value)
{
int i = IndexOfKey(key);
if (i >= 0)
{
value = values[i];
return true;
}
value = default(TV);
return false;
}
// Removes the entry at the given index. The size of the sorted list is
// decreased by one.
public void RemoveAt(int index)
{
if (index < 0 || index >= _size)
throw new ArgumentOutOfRangeException(nameof(index), index, "SR.ArgumentOutOfRange_Index");
_size--;
if (index < _size)
{
Array.Copy(keys, index + 1, keys, index, _size - index);
Array.Copy(values, index + 1, values, index, _size - index);
}
// TODO :
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TKey>())
//{
// keys[_size] = default(TKey);
//}
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TValue>())
//{
// values[_size] = default(TValue);
//}
version++;
}
// Removes an entry from this sorted list. If an entry with the specified
// key exists in the sorted list, it is removed. An ArgumentException is
// thrown if the key is null.
public bool Remove(TK key)
{
int i = IndexOfKey(key);
if (i >= 0)
RemoveAt(i);
return i >= 0;
}
// Sets the capacity of this sorted list to the size of the sorted list.
// This method can be used to minimize a sorted list's memory overhead once
// it is known that no new elements will be added to the sorted list. To
// completely clear a sorted list and release all memory referenced by the
// sorted list, execute the following statements:
//
// SortedList.Clear();
// SortedList.TrimExcess();
public void TrimExcess()
{
int threshold = (int)(((double)keys.Length) * 0.9);
if (_size < threshold)
{
Capacity = _size;
}
}
private static bool IsCompatibleKey(object key)
{
if (key == null)
{
throw new ArgumentNullException(nameof(key));
}
return (key is TK);
}
private struct Enumerator : IEnumerator<KeyValuePair<TK, TV>>, IDictionaryEnumerator
{
private SortedList<TK, TV> _sortedList;
private TK _key;
private TV _value;
private int _index;
private int _version;
private int _getEnumeratorRetType; // What should Enumerator.Current return?
internal const int KeyValuePair = 1;
internal const int DictEntry = 2;
internal Enumerator(SortedList<TK, TV> sortedList, int getEnumeratorRetType)
{
_sortedList = sortedList;
_index = 0;
_version = _sortedList.version;
_getEnumeratorRetType = getEnumeratorRetType;
_key = default(TK);
_value = default(TV);
}
public void Dispose()
{
_index = 0;
_key = default(TK);
_value = default(TV);
}
object IDictionaryEnumerator.Key
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _key;
}
}
public bool MoveNext()
{
if (_version != _sortedList.version) throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
if ((uint)_index < (uint)_sortedList.Count)
{
_key = _sortedList.keys[_index];
_value = _sortedList.values[_index];
_index++;
return true;
}
_index = _sortedList.Count + 1;
_key = default(TK);
_value = default(TV);
return false;
}
DictionaryEntry IDictionaryEnumerator.Entry
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return new DictionaryEntry(_key, _value);
}
}
public KeyValuePair<TK, TV> Current
{
get
{
return new KeyValuePair<TK, TV>(_key, _value);
}
}
object IEnumerator.Current
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
if (_getEnumeratorRetType == DictEntry)
{
return new DictionaryEntry(_key, _value);
}
else
{
return new KeyValuePair<TK, TV>(_key, _value);
}
}
}
object IDictionaryEnumerator.Value
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _value;
}
}
void IEnumerator.Reset()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
_index = 0;
_key = default(TK);
_value = default(TV);
}
}
private sealed class SortedListKeyEnumerator : IEnumerator<TK>, IEnumerator
{
private SortedList<TK, TV> _sortedList;
private int _index;
private int _version;
private TK _currentKey;
internal SortedListKeyEnumerator(SortedList<TK, TV> sortedList)
{
_sortedList = sortedList;
_version = sortedList.version;
}
public void Dispose()
{
_index = 0;
_currentKey = default(TK);
}
public bool MoveNext()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
if ((uint)_index < (uint)_sortedList.Count)
{
_currentKey = _sortedList.keys[_index];
_index++;
return true;
}
_index = _sortedList.Count + 1;
_currentKey = default(TK);
return false;
}
public TK Current
{
get
{
return _currentKey;
}
}
object IEnumerator.Current
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _currentKey;
}
}
void IEnumerator.Reset()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
_index = 0;
_currentKey = default(TK);
}
}
private sealed class SortedListValueEnumerator : IEnumerator<TV>, IEnumerator
{
private SortedList<TK, TV> _sortedList;
private int _index;
private int _version;
private TV _currentValue;
internal SortedListValueEnumerator(SortedList<TK, TV> sortedList)
{
_sortedList = sortedList;
_version = sortedList.version;
}
public void Dispose()
{
_index = 0;
_currentValue = default(TV);
}
public bool MoveNext()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
if ((uint)_index < (uint)_sortedList.Count)
{
_currentValue = _sortedList.values[_index];
_index++;
return true;
}
_index = _sortedList.Count + 1;
_currentValue = default(TV);
return false;
}
public TV Current
{
get
{
return _currentValue;
}
}
object IEnumerator.Current
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _currentValue;
}
}
void IEnumerator.Reset()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
_index = 0;
_currentValue = default(TV);
}
}
[DebuggerDisplay("Count = {Count}")]
[Serializable]
public sealed class KeyList : IList<TK>, ICollection
{
private SortedList<TK, TV> _dict; // Do not rename (binary serialization)
internal KeyList(SortedList<TK, TV> dictionary)
{
_dict = dictionary;
}
public int Count
{
get { return _dict._size; }
}
public bool IsReadOnly
{
get { return true; }
}
bool ICollection.IsSynchronized
{
get { return false; }
}
object ICollection.SyncRoot
{
get { return ((ICollection)_dict).SyncRoot; }
}
public void Add(TK key)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public void Clear()
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public bool Contains(TK key)
{
return _dict.ContainsKey(key);
}
public void CopyTo(TK[] array, int arrayIndex)
{
// defer error checking to Array.Copy
Array.Copy(_dict.keys, 0, array, arrayIndex, _dict.Count);
}
void ICollection.CopyTo(Array array, int arrayIndex)
{
if (array != null && array.Rank != 1)
throw new ArgumentException("SR.Arg_RankMultiDimNotSupported, nameof(array)");
try
{
// defer error checking to Array.Copy
Array.Copy(_dict.keys, 0, array, arrayIndex, _dict.Count);
}
catch (ArrayTypeMismatchException)
{
throw new ArgumentException("SR.Argument_InvalidArrayType, nameof(array)");
}
}
public void Insert(int index, TK value)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public TK this[int index]
{
get
{
return _dict.GetKey(index);
}
set
{
throw new NotSupportedException("SR.NotSupported_KeyCollectionSet");
}
}
public IEnumerator<TK> GetEnumerator()
{
return new SortedListKeyEnumerator(_dict);
}
IEnumerator IEnumerable.GetEnumerator()
{
return new SortedListKeyEnumerator(_dict);
}
public int IndexOf(TK key)
{
if (((object)key) == null)
throw new ArgumentNullException(nameof(key));
int i = Array.BinarySearch<TK>(_dict.keys, 0,
_dict.Count, key, _dict.comparer);
if (i >= 0) return i;
return -1;
}
public bool Remove(TK key)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
// return false;
}
public void RemoveAt(int index)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
}
[DebuggerDisplay("Count = {Count}")]
[Serializable]
public sealed class ValueList : IList<TV>, ICollection
{
private SortedList<TK, TV> _dict; // Do not rename (binary serialization)
internal ValueList(SortedList<TK, TV> dictionary)
{
_dict = dictionary;
}
public int Count
{
get { return _dict._size; }
}
public bool IsReadOnly
{
get { return true; }
}
bool ICollection.IsSynchronized
{
get { return false; }
}
object ICollection.SyncRoot
{
get { return ((ICollection)_dict).SyncRoot; }
}
public void Add(TV key)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public void Clear()
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public bool Contains(TV value)
{
return _dict.ContainsValue(value);
}
public void CopyTo(TV[] array, int arrayIndex)
{
// defer error checking to Array.Copy
Array.Copy(_dict.values, 0, array, arrayIndex, _dict.Count);
}
void ICollection.CopyTo(Array array, int index)
{
if (array != null && array.Rank != 1)
throw new ArgumentException("SR.Arg_RankMultiDimNotSupported, nameof(array)");
try
{
// defer error checking to Array.Copy
Array.Copy(_dict.values, 0, array, index, _dict.Count);
}
catch (ArrayTypeMismatchException)
{
throw new ArgumentException("SR.Argument_InvalidArrayType, nameof(array)");
}
}
public void Insert(int index, TV value)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public TV this[int index]
{
get
{
return _dict.GetByIndex(index);
}
set
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
}
public IEnumerator<TV> GetEnumerator()
{
return new SortedListValueEnumerator(_dict);
}
IEnumerator IEnumerable.GetEnumerator()
{
return new SortedListValueEnumerator(_dict);
}
public int IndexOf(TV value)
{
return Array.IndexOf(_dict.values, value, 0, _dict.Count);
}
public bool Remove(TV value)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
// return false;
}
public void RemoveAt(int index)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
}
}
public static void Main()
{
var sortedList = new SortedList<int, int>();
var msg = "Hello, World! " + sortedList.Count();
Console.WriteLine(msg);
// Try other C# Samples from
// the Button in the top right.
}
}
```
### Expected Result
No runtime error
### Actual Result
System.Exception: TypeError: Cannot read property '$interfaces' of undefined
at Object.createInheritors (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:36504)
at Function.i.$staticInit (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:40326)
at Object.init (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:41114)
at Object.Bridge.assembly (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:41809)
at https://deck.net/RunHandler.ashx?h=1760218166:4:8 | 1.0 | SortedList implementation causes Javascript runtime error - I copied dotnet source for sortedlist implementation, and removed the resource string. The resulting code compiles fine, but Javascript execution fails.
I tracked the issue down to the sub-class
private struct Enumerator : IEnumerator<KeyValuePair<TK, TV>>, IDictionaryEnumerator
If I remove the Enumerator class, runtime error goes away but of course the sortedlist does not work without the enumerator.
Further, If I leave the sub-class, and remove IDictionaryEnumerator, then it works.
### Steps To Reproduce
https://deck.net/94fd47356bddf4c009b256a472561736
```csharp
public class Program
{
public class SortedList<TK, TV> :
IDictionary<TK, TV>
{
private TK[] keys; // Do not rename (binary serialization)
private TV[] values; // Do not rename (binary serialization)
private int _size; // Do not rename (binary serialization)
private int version; // Do not rename (binary serialization)
private IComparer<TK> comparer; // Do not rename (binary serialization)
private KeyList keyList; // Do not rename (binary serialization)
private ValueList valueList; // Do not rename (binary serialization)
[NonSerialized]
private object _syncRoot;
private const int DefaultCapacity = 4;
// Constructs a new sorted list. The sorted list is initially empty and has
// a capacity of zero. Upon adding the first element to the sorted list the
// capacity is increased to DefaultCapacity, and then increased in multiples of two as
// required. The elements of the sorted list are ordered according to the
// IComparable interface, which must be implemented by the keys of
// all entries added to the sorted list.
public SortedList()
{
keys = new TK[] { };
values = new TV[] { };
_size = 0;
comparer = Comparer<TK>.Default;
}
// Constructs a new sorted list. The sorted list is initially empty and has
// a capacity of zero. Upon adding the first element to the sorted list the
// capacity is increased to 16, and then increased in multiples of two as
// required. The elements of the sorted list are ordered according to the
// IComparable interface, which must be implemented by the keys of
// all entries added to the sorted list.
//
public SortedList(int capacity)
{
if (capacity < 0)
throw new ArgumentOutOfRangeException(nameof(capacity), capacity, "SR.ArgumentOutOfRange_NeedNonNegNum");
keys = new TK[capacity];
values = new TV[capacity];
comparer = Comparer<TK>.Default;
}
// Constructs a new sorted list with a given IComparer
// implementation. The sorted list is initially empty and has a capacity of
// zero. Upon adding the first element to the sorted list the capacity is
// increased to 16, and then increased in multiples of two as required. The
// elements of the sorted list are ordered according to the given
// IComparer implementation. If comparer is null, the
// elements are compared to each other using the IComparable
// interface, which in that case must be implemented by the keys of all
// entries added to the sorted list.
//
public SortedList(IComparer<TK> comparer)
: this()
{
if (comparer != null)
{
this.comparer = comparer;
}
}
// Constructs a new sorted dictionary with a given IComparer
// implementation and a given initial capacity. The sorted list is
// initially empty, but will have room for the given number of elements
// before any reallocations are required. The elements of the sorted list
// are ordered according to the given IComparer implementation. If
// comparer is null, the elements are compared to each other using
// the IComparable interface, which in that case must be implemented
// by the keys of all entries added to the sorted list.
//
public SortedList(int capacity, IComparer<TK> comparer)
: this(comparer)
{
Capacity = capacity;
}
// Constructs a new sorted list containing a copy of the entries in the
// given dictionary. The elements of the sorted list are ordered according
// to the IComparable interface, which must be implemented by the
// keys of all entries in the given dictionary as well as keys
// subsequently added to the sorted list.
//
public SortedList(IDictionary<TK, TV> dictionary)
: this(dictionary, null)
{
}
// Constructs a new sorted list containing a copy of the entries in the
// given dictionary. The elements of the sorted list are ordered according
// to the given IComparer implementation. If comparer is
// null, the elements are compared to each other using the
// IComparable interface, which in that case must be implemented
// by the keys of all entries in the given dictionary as well as keys
// subsequently added to the sorted list.
//
public SortedList(IDictionary<TK, TV> dictionary, IComparer<TK> comparer)
: this((dictionary != null ? dictionary.Count : 0), comparer)
{
if (dictionary == null)
throw new ArgumentNullException(nameof(dictionary));
int count = dictionary.Count;
if (count != 0)
{
TK[] keys = this.keys;
dictionary.Keys.CopyTo(keys, 0);
dictionary.Values.CopyTo(values, 0);
Debug.Assert(count == this.keys.Length);
if (count > 1)
{
comparer = Comparer; // obtain default if this is null.
keys = dictionary.OrderBy(x => x.Key, comparer).Select(x => x.Key).ToArray();
for (int i = 1; i != keys.Length; ++i)
{
if (comparer.Compare(keys[i - 1], keys[i]) == 0)
{
throw new ArgumentException("SR.Format(SR.Argument_AddingDuplicate, keys[i])");
}
}
}
}
_size = count;
}
// Adds an entry with the given key and value to this sorted list. An
// ArgumentException is thrown if the key is already present in the sorted list.
//
public void Add(TK key, TV value)
{
if (key == null) throw new ArgumentNullException(nameof(key));
int i = Array.BinarySearch<TK>(keys, 0, _size, key, comparer);
if (i >= 0)
throw new ArgumentException("SR.Format(SR.Argument_AddingDuplicate, key), nameof(key)");
Insert(~i, key, value);
}
void ICollection<KeyValuePair<TK, TV>>.Add(KeyValuePair<TK, TV> keyValuePair)
{
Add(keyValuePair.Key, keyValuePair.Value);
}
bool ICollection<KeyValuePair<TK, TV>>.Contains(KeyValuePair<TK, TV> keyValuePair)
{
int index = IndexOfKey(keyValuePair.Key);
if (index >= 0 && EqualityComparer<TV>.Default.Equals(values[index], keyValuePair.Value))
{
return true;
}
return false;
}
bool ICollection<KeyValuePair<TK, TV>>.Remove(KeyValuePair<TK, TV> keyValuePair)
{
int index = IndexOfKey(keyValuePair.Key);
if (index >= 0 && EqualityComparer<TV>.Default.Equals(values[index], keyValuePair.Value))
{
RemoveAt(index);
return true;
}
return false;
}
// Returns the capacity of this sorted list. The capacity of a sorted list
// represents the allocated length of the internal arrays used to store the
// keys and values of the list, and thus also indicates the maximum number
// of entries the list can contain before a reallocation of the internal
// arrays is required.
//
public int Capacity
{
get
{
return keys.Length;
}
set
{
if (value != keys.Length)
{
if (value < _size)
{
throw new ArgumentOutOfRangeException(nameof(value), value, "SR.ArgumentOutOfRange_SmallCapacity");
}
if (value > 0)
{
TK[] newKeys = new TK[value];
TV[] newValues = new TV[value];
if (_size > 0)
{
Array.Copy(keys, 0, newKeys, 0, _size);
Array.Copy(values, 0, newValues, 0, _size);
}
keys = newKeys;
values = newValues;
}
else
{
keys = new TK[] { };
values = new TV[] { };
}
}
}
}
public IComparer<TK> Comparer
{
get
{
return comparer;
}
}
// Returns the number of entries in this sorted list.
public int Count
{
get
{
return _size;
}
}
// Returns a collection representing the keys of this sorted list. This
// method returns the same object as GetKeyList, but typed as an
// ICollection instead of an IList.
public IList<TK> Keys
{
get
{
return GetKeyListHelper();
}
}
ICollection<TK> IDictionary<TK, TV>.Keys
{
get
{
return GetKeyListHelper();
}
}
// Returns a collection representing the values of this sorted list. This
// method returns the same object as GetValueList, but typed as an
// ICollection instead of an IList.
//
public IList<TV> Values
{
get
{
return GetValueListHelper();
}
}
ICollection<TV> IDictionary<TK, TV>.Values
{
get
{
return GetValueListHelper();
}
}
private KeyList GetKeyListHelper()
{
if (keyList == null)
keyList = new KeyList(this);
return keyList;
}
private ValueList GetValueListHelper()
{
if (valueList == null)
valueList = new ValueList(this);
return valueList;
}
bool ICollection<KeyValuePair<TK, TV>>.IsReadOnly
{
get { return false; }
}
// Removes all entries from this sorted list.
public void Clear()
{
// clear does not change the capacity
version++;
// TODO:
// Don't need to doc this but we clear the elements so that the gc can reclaim the references.
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TKey>())
//{
// Array.Clear(keys, 0, _size);
//}
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TValue>())
//{
// Array.Clear(values, 0, _size);
//}
_size = 0;
}
// Checks if this sorted list contains an entry with the given key.
public bool ContainsKey(TK key)
{
return IndexOfKey(key) >= 0;
}
// Checks if this sorted list contains an entry with the given value. The
// values of the entries of the sorted list are compared to the given value
// using the Object.Equals method. This method performs a linear
// search and is substantially slower than the Contains
// method.
public bool ContainsValue(TV value)
{
return IndexOfValue(value) >= 0;
}
// Copies the values in this SortedList to an array.
void ICollection<KeyValuePair<TK, TV>>.CopyTo(KeyValuePair<TK, TV>[] array, int arrayIndex)
{
if (array == null)
{
throw new ArgumentNullException(nameof(array));
}
if (arrayIndex < 0 || arrayIndex > array.Length)
{
throw new ArgumentOutOfRangeException(nameof(arrayIndex), arrayIndex, "SR.ArgumentOutOfRange_Index");
}
if (array.Length - arrayIndex < Count)
{
throw new ArgumentException("SR.Arg_ArrayPlusOffTooSmall");
}
for (int i = 0; i < Count; i++)
{
KeyValuePair<TK, TV> entry = new KeyValuePair<TK, TV>(keys[i], values[i]);
array[arrayIndex + i] = entry;
}
}
private const int MaxArrayLength = 0X7FEFFFFF;
// Ensures that the capacity of this sorted list is at least the given
// minimum value. If the current capacity of the list is less than
// min, the capacity is increased to twice the current capacity or
// to min, whichever is larger.
private void EnsureCapacity(int min)
{
int newCapacity = keys.Length == 0 ? DefaultCapacity : keys.Length * 2;
// Allow the list to grow to maximum possible capacity (~2G elements) before encountering overflow.
// Note that this check works even when _items.Length overflowed thanks to the (uint) cast
if ((uint)newCapacity > MaxArrayLength) newCapacity = MaxArrayLength;
if (newCapacity < min) newCapacity = min;
Capacity = newCapacity;
}
// Returns the value of the entry at the given index.
private TV GetByIndex(int index)
{
if (index < 0 || index >= _size)
throw new ArgumentOutOfRangeException(nameof(index), index, "SR.ArgumentOutOfRange_Index");
return values[index];
}
public IEnumerator<KeyValuePair<TK, TV>> GetEnumerator()
{
return new Enumerator(this, Enumerator.KeyValuePair);
}
IEnumerator<KeyValuePair<TK, TV>> IEnumerable<KeyValuePair<TK, TV>>.GetEnumerator()
{
return new Enumerator(this, Enumerator.KeyValuePair);
}
IEnumerator IEnumerable.GetEnumerator()
{
return new Enumerator(this, Enumerator.KeyValuePair);
}
// Returns the key of the entry at the given index.
private TK GetKey(int index)
{
if (index < 0 || index >= _size)
throw new ArgumentOutOfRangeException(nameof(index), index, "SR.ArgumentOutOfRange_Index");
return keys[index];
}
// Returns the value associated with the given key. If an entry with the
// given key is not found, the returned value is null.
public TV this[TK key]
{
get
{
int i = IndexOfKey(key);
if (i >= 0)
return values[i];
throw new KeyNotFoundException("SR.Format(SR.Arg_KeyNotFoundWithKey, key.ToString())");
}
set
{
if (((object)key) == null) throw new ArgumentNullException(nameof(key));
int i = Array.BinarySearch<TK>(keys, 0, _size, key, comparer);
if (i >= 0)
{
values[i] = value;
version++;
return;
}
Insert(~i, key, value);
}
}
// Returns the index of the entry with a given key in this sorted list. The
// key is located through a binary search, and thus the average execution
// time of this method is proportional to Log2(size), where
// size is the size of this sorted list. The returned value is -1 if
// the given key does not occur in this sorted list. Null is an invalid
// key value.
public int IndexOfKey(TK key)
{
if (key == null)
throw new ArgumentNullException(nameof(key));
int ret = Array.BinarySearch<TK>(keys, 0, _size, key, comparer);
return ret >= 0 ? ret : -1;
}
// Returns the index of the first occurrence of an entry with a given value
// in this sorted list. The entry is located through a linear search, and
// thus the average execution time of this method is proportional to the
// size of this sorted list. The elements of the list are compared to the
// given value using the Object.Equals method.
public int IndexOfValue(TV value)
{
return Array.IndexOf(values, value, 0, _size);
}
// Inserts an entry with a given key and value at a given index.
private void Insert(int index, TK key, TV value)
{
if (_size == keys.Length) EnsureCapacity(_size + 1);
if (index < _size)
{
Array.Copy(keys, index, keys, index + 1, _size - index);
Array.Copy(values, index, values, index + 1, _size - index);
}
keys[index] = key;
values[index] = value;
_size++;
version++;
}
public bool TryGetValue(TK key, out TV value)
{
int i = IndexOfKey(key);
if (i >= 0)
{
value = values[i];
return true;
}
value = default(TV);
return false;
}
// Removes the entry at the given index. The size of the sorted list is
// decreased by one.
public void RemoveAt(int index)
{
if (index < 0 || index >= _size)
throw new ArgumentOutOfRangeException(nameof(index), index, "SR.ArgumentOutOfRange_Index");
_size--;
if (index < _size)
{
Array.Copy(keys, index + 1, keys, index, _size - index);
Array.Copy(values, index + 1, values, index, _size - index);
}
// TODO :
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TKey>())
//{
// keys[_size] = default(TKey);
//}
//if (RuntimeHelpers.IsReferenceOrContainsReferences<TValue>())
//{
// values[_size] = default(TValue);
//}
version++;
}
// Removes an entry from this sorted list. If an entry with the specified
// key exists in the sorted list, it is removed. An ArgumentException is
// thrown if the key is null.
public bool Remove(TK key)
{
int i = IndexOfKey(key);
if (i >= 0)
RemoveAt(i);
return i >= 0;
}
// Sets the capacity of this sorted list to the size of the sorted list.
// This method can be used to minimize a sorted list's memory overhead once
// it is known that no new elements will be added to the sorted list. To
// completely clear a sorted list and release all memory referenced by the
// sorted list, execute the following statements:
//
// SortedList.Clear();
// SortedList.TrimExcess();
public void TrimExcess()
{
int threshold = (int)(((double)keys.Length) * 0.9);
if (_size < threshold)
{
Capacity = _size;
}
}
private static bool IsCompatibleKey(object key)
{
if (key == null)
{
throw new ArgumentNullException(nameof(key));
}
return (key is TK);
}
private struct Enumerator : IEnumerator<KeyValuePair<TK, TV>>, IDictionaryEnumerator
{
private SortedList<TK, TV> _sortedList;
private TK _key;
private TV _value;
private int _index;
private int _version;
private int _getEnumeratorRetType; // What should Enumerator.Current return?
internal const int KeyValuePair = 1;
internal const int DictEntry = 2;
internal Enumerator(SortedList<TK, TV> sortedList, int getEnumeratorRetType)
{
_sortedList = sortedList;
_index = 0;
_version = _sortedList.version;
_getEnumeratorRetType = getEnumeratorRetType;
_key = default(TK);
_value = default(TV);
}
public void Dispose()
{
_index = 0;
_key = default(TK);
_value = default(TV);
}
object IDictionaryEnumerator.Key
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _key;
}
}
public bool MoveNext()
{
if (_version != _sortedList.version) throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
if ((uint)_index < (uint)_sortedList.Count)
{
_key = _sortedList.keys[_index];
_value = _sortedList.values[_index];
_index++;
return true;
}
_index = _sortedList.Count + 1;
_key = default(TK);
_value = default(TV);
return false;
}
DictionaryEntry IDictionaryEnumerator.Entry
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return new DictionaryEntry(_key, _value);
}
}
public KeyValuePair<TK, TV> Current
{
get
{
return new KeyValuePair<TK, TV>(_key, _value);
}
}
object IEnumerator.Current
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
if (_getEnumeratorRetType == DictEntry)
{
return new DictionaryEntry(_key, _value);
}
else
{
return new KeyValuePair<TK, TV>(_key, _value);
}
}
}
object IDictionaryEnumerator.Value
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _value;
}
}
void IEnumerator.Reset()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
_index = 0;
_key = default(TK);
_value = default(TV);
}
}
private sealed class SortedListKeyEnumerator : IEnumerator<TK>, IEnumerator
{
private SortedList<TK, TV> _sortedList;
private int _index;
private int _version;
private TK _currentKey;
internal SortedListKeyEnumerator(SortedList<TK, TV> sortedList)
{
_sortedList = sortedList;
_version = sortedList.version;
}
public void Dispose()
{
_index = 0;
_currentKey = default(TK);
}
public bool MoveNext()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
if ((uint)_index < (uint)_sortedList.Count)
{
_currentKey = _sortedList.keys[_index];
_index++;
return true;
}
_index = _sortedList.Count + 1;
_currentKey = default(TK);
return false;
}
public TK Current
{
get
{
return _currentKey;
}
}
object IEnumerator.Current
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _currentKey;
}
}
void IEnumerator.Reset()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
_index = 0;
_currentKey = default(TK);
}
}
private sealed class SortedListValueEnumerator : IEnumerator<TV>, IEnumerator
{
private SortedList<TK, TV> _sortedList;
private int _index;
private int _version;
private TV _currentValue;
internal SortedListValueEnumerator(SortedList<TK, TV> sortedList)
{
_sortedList = sortedList;
_version = sortedList.version;
}
public void Dispose()
{
_index = 0;
_currentValue = default(TV);
}
public bool MoveNext()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
if ((uint)_index < (uint)_sortedList.Count)
{
_currentValue = _sortedList.values[_index];
_index++;
return true;
}
_index = _sortedList.Count + 1;
_currentValue = default(TV);
return false;
}
public TV Current
{
get
{
return _currentValue;
}
}
object IEnumerator.Current
{
get
{
if (_index == 0 || (_index == _sortedList.Count + 1))
{
throw new InvalidOperationException("SR.InvalidOperation_EnumOpCantHappen");
}
return _currentValue;
}
}
void IEnumerator.Reset()
{
if (_version != _sortedList.version)
{
throw new InvalidOperationException("SR.InvalidOperation_EnumFailedVersion");
}
_index = 0;
_currentValue = default(TV);
}
}
[DebuggerDisplay("Count = {Count}")]
[Serializable]
public sealed class KeyList : IList<TK>, ICollection
{
private SortedList<TK, TV> _dict; // Do not rename (binary serialization)
internal KeyList(SortedList<TK, TV> dictionary)
{
_dict = dictionary;
}
public int Count
{
get { return _dict._size; }
}
public bool IsReadOnly
{
get { return true; }
}
bool ICollection.IsSynchronized
{
get { return false; }
}
object ICollection.SyncRoot
{
get { return ((ICollection)_dict).SyncRoot; }
}
public void Add(TK key)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public void Clear()
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public bool Contains(TK key)
{
return _dict.ContainsKey(key);
}
public void CopyTo(TK[] array, int arrayIndex)
{
// defer error checking to Array.Copy
Array.Copy(_dict.keys, 0, array, arrayIndex, _dict.Count);
}
void ICollection.CopyTo(Array array, int arrayIndex)
{
if (array != null && array.Rank != 1)
throw new ArgumentException("SR.Arg_RankMultiDimNotSupported, nameof(array)");
try
{
// defer error checking to Array.Copy
Array.Copy(_dict.keys, 0, array, arrayIndex, _dict.Count);
}
catch (ArrayTypeMismatchException)
{
throw new ArgumentException("SR.Argument_InvalidArrayType, nameof(array)");
}
}
public void Insert(int index, TK value)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public TK this[int index]
{
get
{
return _dict.GetKey(index);
}
set
{
throw new NotSupportedException("SR.NotSupported_KeyCollectionSet");
}
}
public IEnumerator<TK> GetEnumerator()
{
return new SortedListKeyEnumerator(_dict);
}
IEnumerator IEnumerable.GetEnumerator()
{
return new SortedListKeyEnumerator(_dict);
}
public int IndexOf(TK key)
{
if (((object)key) == null)
throw new ArgumentNullException(nameof(key));
int i = Array.BinarySearch<TK>(_dict.keys, 0,
_dict.Count, key, _dict.comparer);
if (i >= 0) return i;
return -1;
}
public bool Remove(TK key)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
// return false;
}
public void RemoveAt(int index)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
}
[DebuggerDisplay("Count = {Count}")]
[Serializable]
public sealed class ValueList : IList<TV>, ICollection
{
private SortedList<TK, TV> _dict; // Do not rename (binary serialization)
internal ValueList(SortedList<TK, TV> dictionary)
{
_dict = dictionary;
}
public int Count
{
get { return _dict._size; }
}
public bool IsReadOnly
{
get { return true; }
}
bool ICollection.IsSynchronized
{
get { return false; }
}
object ICollection.SyncRoot
{
get { return ((ICollection)_dict).SyncRoot; }
}
public void Add(TV key)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public void Clear()
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public bool Contains(TV value)
{
return _dict.ContainsValue(value);
}
public void CopyTo(TV[] array, int arrayIndex)
{
// defer error checking to Array.Copy
Array.Copy(_dict.values, 0, array, arrayIndex, _dict.Count);
}
void ICollection.CopyTo(Array array, int index)
{
if (array != null && array.Rank != 1)
throw new ArgumentException("SR.Arg_RankMultiDimNotSupported, nameof(array)");
try
{
// defer error checking to Array.Copy
Array.Copy(_dict.values, 0, array, index, _dict.Count);
}
catch (ArrayTypeMismatchException)
{
throw new ArgumentException("SR.Argument_InvalidArrayType, nameof(array)");
}
}
public void Insert(int index, TV value)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
public TV this[int index]
{
get
{
return _dict.GetByIndex(index);
}
set
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
}
public IEnumerator<TV> GetEnumerator()
{
return new SortedListValueEnumerator(_dict);
}
IEnumerator IEnumerable.GetEnumerator()
{
return new SortedListValueEnumerator(_dict);
}
public int IndexOf(TV value)
{
return Array.IndexOf(_dict.values, value, 0, _dict.Count);
}
public bool Remove(TV value)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
// return false;
}
public void RemoveAt(int index)
{
throw new NotSupportedException("SR.NotSupported_SortedListNestedWrite");
}
}
}
public static void Main()
{
var sortedList = new SortedList<int, int>();
var msg = "Hello, World! " + sortedList.Count();
Console.WriteLine(msg);
// Try other C# Samples from
// the Button in the top right.
}
}
```
### Expected Result
No runtime error
### Actual Result
System.Exception: TypeError: Cannot read property '$interfaces' of undefined
at Object.createInheritors (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:36504)
at Function.i.$staticInit (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:40326)
at Object.init (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:41114)
at Object.Bridge.assembly (https://deck.net/resources/js/bridge/bridge.min.js?17.1.0-201805031411:7:41809)
at https://deck.net/RunHandler.ashx?h=1760218166:4:8 | defect | sortedlist implementation causes javascript runtime error i copied dotnet source for sortedlist implementation and removed the resource string the resulting code compiles fine but javascript execution fails i tracked the issue down to the sub class private struct enumerator ienumerator idictionaryenumerator if i remove the enumerator class runtime error goes away but of course the sortedlist does not work without the enumerator further if i leave the sub class and remove idictionaryenumerator then it works steps to reproduce csharp public class program public class sortedlist idictionary private tk keys do not rename binary serialization private tv values do not rename binary serialization private int size do not rename binary serialization private int version do not rename binary serialization private icomparer comparer do not rename binary serialization private keylist keylist do not rename binary serialization private valuelist valuelist do not rename binary serialization private object syncroot private const int defaultcapacity constructs a new sorted list the sorted list is initially empty and has a capacity of zero upon adding the first element to the sorted list the capacity is increased to defaultcapacity and then increased in multiples of two as required the elements of the sorted list are ordered according to the icomparable interface which must be implemented by the keys of all entries added to the sorted list public sortedlist keys new tk values new tv size comparer comparer default constructs a new sorted list the sorted list is initially empty and has a capacity of zero upon adding the first element to the sorted list the capacity is increased to and then increased in multiples of two as required the elements of the sorted list are ordered according to the icomparable interface which must be implemented by the keys of all entries added to the sorted list public sortedlist int capacity if capacity throw new argumentoutofrangeexception nameof capacity capacity sr argumentoutofrange neednonnegnum keys new tk values new tv comparer comparer default constructs a new sorted list with a given icomparer implementation the sorted list is initially empty and has a capacity of zero upon adding the first element to the sorted list the capacity is increased to and then increased in multiples of two as required the elements of the sorted list are ordered according to the given icomparer implementation if comparer is null the elements are compared to each other using the icomparable interface which in that case must be implemented by the keys of all entries added to the sorted list public sortedlist icomparer comparer this if comparer null this comparer comparer constructs a new sorted dictionary with a given icomparer implementation and a given initial capacity the sorted list is initially empty but will have room for the given number of elements before any reallocations are required the elements of the sorted list are ordered according to the given icomparer implementation if comparer is null the elements are compared to each other using the icomparable interface which in that case must be implemented by the keys of all entries added to the sorted list public sortedlist int capacity icomparer comparer this comparer capacity capacity constructs a new sorted list containing a copy of the entries in the given dictionary the elements of the sorted list are ordered according to the icomparable interface which must be implemented by the keys of all entries in the given dictionary as well as keys subsequently added to the sorted list public sortedlist idictionary dictionary this dictionary null constructs a new sorted list containing a copy of the entries in the given dictionary the elements of the sorted list are ordered according to the given icomparer implementation if comparer is null the elements are compared to each other using the icomparable interface which in that case must be implemented by the keys of all entries in the given dictionary as well as keys subsequently added to the sorted list public sortedlist idictionary dictionary icomparer comparer this dictionary null dictionary count comparer if dictionary null throw new argumentnullexception nameof dictionary int count dictionary count if count tk keys this keys dictionary keys copyto keys dictionary values copyto values debug assert count this keys length if count comparer comparer obtain default if this is null keys dictionary orderby x x key comparer select x x key toarray for int i i keys length i if comparer compare keys keys throw new argumentexception sr format sr argument addingduplicate keys size count adds an entry with the given key and value to this sorted list an argumentexception is thrown if the key is already present in the sorted list public void add tk key tv value if key null throw new argumentnullexception nameof key int i array binarysearch keys size key comparer if i throw new argumentexception sr format sr argument addingduplicate key nameof key insert i key value void icollection add keyvaluepair keyvaluepair add keyvaluepair key keyvaluepair value bool icollection contains keyvaluepair keyvaluepair int index indexofkey keyvaluepair key if index equalitycomparer default equals values keyvaluepair value return true return false bool icollection remove keyvaluepair keyvaluepair int index indexofkey keyvaluepair key if index equalitycomparer default equals values keyvaluepair value removeat index return true return false returns the capacity of this sorted list the capacity of a sorted list represents the allocated length of the internal arrays used to store the keys and values of the list and thus also indicates the maximum number of entries the list can contain before a reallocation of the internal arrays is required public int capacity get return keys length set if value keys length if value size throw new argumentoutofrangeexception nameof value value sr argumentoutofrange smallcapacity if value tk newkeys new tk tv newvalues new tv if size array copy keys newkeys size array copy values newvalues size keys newkeys values newvalues else keys new tk values new tv public icomparer comparer get return comparer returns the number of entries in this sorted list public int count get return size returns a collection representing the keys of this sorted list this method returns the same object as getkeylist but typed as an icollection instead of an ilist public ilist keys get return getkeylisthelper icollection idictionary keys get return getkeylisthelper returns a collection representing the values of this sorted list this method returns the same object as getvaluelist but typed as an icollection instead of an ilist public ilist values get return getvaluelisthelper icollection idictionary values get return getvaluelisthelper private keylist getkeylisthelper if keylist null keylist new keylist this return keylist private valuelist getvaluelisthelper if valuelist null valuelist new valuelist this return valuelist bool icollection isreadonly get return false removes all entries from this sorted list public void clear clear does not change the capacity version todo don t need to doc this but we clear the elements so that the gc can reclaim the references if runtimehelpers isreferenceorcontainsreferences array clear keys size if runtimehelpers isreferenceorcontainsreferences array clear values size size checks if this sorted list contains an entry with the given key public bool containskey tk key return indexofkey key checks if this sorted list contains an entry with the given value the values of the entries of the sorted list are compared to the given value using the object equals method this method performs a linear search and is substantially slower than the contains method public bool containsvalue tv value return indexofvalue value copies the values in this sortedlist to an array void icollection copyto keyvaluepair array int arrayindex if array null throw new argumentnullexception nameof array if arrayindex array length throw new argumentoutofrangeexception nameof arrayindex arrayindex sr argumentoutofrange index if array length arrayindex count throw new argumentexception sr arg arrayplusofftoosmall for int i i count i keyvaluepair entry new keyvaluepair keys values array entry private const int maxarraylength ensures that the capacity of this sorted list is at least the given minimum value if the current capacity of the list is less than min the capacity is increased to twice the current capacity or to min whichever is larger private void ensurecapacity int min int newcapacity keys length defaultcapacity keys length allow the list to grow to maximum possible capacity elements before encountering overflow note that this check works even when items length overflowed thanks to the uint cast if uint newcapacity maxarraylength newcapacity maxarraylength if newcapacity min newcapacity min capacity newcapacity returns the value of the entry at the given index private tv getbyindex int index if index size throw new argumentoutofrangeexception nameof index index sr argumentoutofrange index return values public ienumerator getenumerator return new enumerator this enumerator keyvaluepair ienumerator ienumerable getenumerator return new enumerator this enumerator keyvaluepair ienumerator ienumerable getenumerator return new enumerator this enumerator keyvaluepair returns the key of the entry at the given index private tk getkey int index if index size throw new argumentoutofrangeexception nameof index index sr argumentoutofrange index return keys returns the value associated with the given key if an entry with the given key is not found the returned value is null public tv this get int i indexofkey key if i return values throw new keynotfoundexception sr format sr arg keynotfoundwithkey key tostring set if object key null throw new argumentnullexception nameof key int i array binarysearch keys size key comparer if i values value version return insert i key value returns the index of the entry with a given key in this sorted list the key is located through a binary search and thus the average execution time of this method is proportional to size where size is the size of this sorted list the returned value is if the given key does not occur in this sorted list null is an invalid key value public int indexofkey tk key if key null throw new argumentnullexception nameof key int ret array binarysearch keys size key comparer return ret ret returns the index of the first occurrence of an entry with a given value in this sorted list the entry is located through a linear search and thus the average execution time of this method is proportional to the size of this sorted list the elements of the list are compared to the given value using the object equals method public int indexofvalue tv value return array indexof values value size inserts an entry with a given key and value at a given index private void insert int index tk key tv value if size keys length ensurecapacity size if index size array copy keys index keys index size index array copy values index values index size index keys key values value size version public bool trygetvalue tk key out tv value int i indexofkey key if i value values return true value default tv return false removes the entry at the given index the size of the sorted list is decreased by one public void removeat int index if index size throw new argumentoutofrangeexception nameof index index sr argumentoutofrange index size if index size array copy keys index keys index size index array copy values index values index size index todo if runtimehelpers isreferenceorcontainsreferences keys default tkey if runtimehelpers isreferenceorcontainsreferences values default tvalue version removes an entry from this sorted list if an entry with the specified key exists in the sorted list it is removed an argumentexception is thrown if the key is null public bool remove tk key int i indexofkey key if i removeat i return i sets the capacity of this sorted list to the size of the sorted list this method can be used to minimize a sorted list s memory overhead once it is known that no new elements will be added to the sorted list to completely clear a sorted list and release all memory referenced by the sorted list execute the following statements sortedlist clear sortedlist trimexcess public void trimexcess int threshold int double keys length if size threshold capacity size private static bool iscompatiblekey object key if key null throw new argumentnullexception nameof key return key is tk private struct enumerator ienumerator idictionaryenumerator private sortedlist sortedlist private tk key private tv value private int index private int version private int getenumeratorrettype what should enumerator current return internal const int keyvaluepair internal const int dictentry internal enumerator sortedlist sortedlist int getenumeratorrettype sortedlist sortedlist index version sortedlist version getenumeratorrettype getenumeratorrettype key default tk value default tv public void dispose index key default tk value default tv object idictionaryenumerator key get if index index sortedlist count throw new invalidoperationexception sr invalidoperation enumopcanthappen return key public bool movenext if version sortedlist version throw new invalidoperationexception sr invalidoperation enumfailedversion if uint index uint sortedlist count key sortedlist keys value sortedlist values index return true index sortedlist count key default tk value default tv return false dictionaryentry idictionaryenumerator entry get if index index sortedlist count throw new invalidoperationexception sr invalidoperation enumopcanthappen return new dictionaryentry key value public keyvaluepair current get return new keyvaluepair key value object ienumerator current get if index index sortedlist count throw new invalidoperationexception sr invalidoperation enumopcanthappen if getenumeratorrettype dictentry return new dictionaryentry key value else return new keyvaluepair key value object idictionaryenumerator value get if index index sortedlist count throw new invalidoperationexception sr invalidoperation enumopcanthappen return value void ienumerator reset if version sortedlist version throw new invalidoperationexception sr invalidoperation enumfailedversion index key default tk value default tv private sealed class sortedlistkeyenumerator ienumerator ienumerator private sortedlist sortedlist private int index private int version private tk currentkey internal sortedlistkeyenumerator sortedlist sortedlist sortedlist sortedlist version sortedlist version public void dispose index currentkey default tk public bool movenext if version sortedlist version throw new invalidoperationexception sr invalidoperation enumfailedversion if uint index uint sortedlist count currentkey sortedlist keys index return true index sortedlist count currentkey default tk return false public tk current get return currentkey object ienumerator current get if index index sortedlist count throw new invalidoperationexception sr invalidoperation enumopcanthappen return currentkey void ienumerator reset if version sortedlist version throw new invalidoperationexception sr invalidoperation enumfailedversion index currentkey default tk private sealed class sortedlistvalueenumerator ienumerator ienumerator private sortedlist sortedlist private int index private int version private tv currentvalue internal sortedlistvalueenumerator sortedlist sortedlist sortedlist sortedlist version sortedlist version public void dispose index currentvalue default tv public bool movenext if version sortedlist version throw new invalidoperationexception sr invalidoperation enumfailedversion if uint index uint sortedlist count currentvalue sortedlist values index return true index sortedlist count currentvalue default tv return false public tv current get return currentvalue object ienumerator current get if index index sortedlist count throw new invalidoperationexception sr invalidoperation enumopcanthappen return currentvalue void ienumerator reset if version sortedlist version throw new invalidoperationexception sr invalidoperation enumfailedversion index currentvalue default tv public sealed class keylist ilist icollection private sortedlist dict do not rename binary serialization internal keylist sortedlist dictionary dict dictionary public int count get return dict size public bool isreadonly get return true bool icollection issynchronized get return false object icollection syncroot get return icollection dict syncroot public void add tk key throw new notsupportedexception sr notsupported sortedlistnestedwrite public void clear throw new notsupportedexception sr notsupported sortedlistnestedwrite public bool contains tk key return dict containskey key public void copyto tk array int arrayindex defer error checking to array copy array copy dict keys array arrayindex dict count void icollection copyto array array int arrayindex if array null array rank throw new argumentexception sr arg rankmultidimnotsupported nameof array try defer error checking to array copy array copy dict keys array arrayindex dict count catch arraytypemismatchexception throw new argumentexception sr argument invalidarraytype nameof array public void insert int index tk value throw new notsupportedexception sr notsupported sortedlistnestedwrite public tk this get return dict getkey index set throw new notsupportedexception sr notsupported keycollectionset public ienumerator getenumerator return new sortedlistkeyenumerator dict ienumerator ienumerable getenumerator return new sortedlistkeyenumerator dict public int indexof tk key if object key null throw new argumentnullexception nameof key int i array binarysearch dict keys dict count key dict comparer if i return i return public bool remove tk key throw new notsupportedexception sr notsupported sortedlistnestedwrite return false public void removeat int index throw new notsupportedexception sr notsupported sortedlistnestedwrite public sealed class valuelist ilist icollection private sortedlist dict do not rename binary serialization internal valuelist sortedlist dictionary dict dictionary public int count get return dict size public bool isreadonly get return true bool icollection issynchronized get return false object icollection syncroot get return icollection dict syncroot public void add tv key throw new notsupportedexception sr notsupported sortedlistnestedwrite public void clear throw new notsupportedexception sr notsupported sortedlistnestedwrite public bool contains tv value return dict containsvalue value public void copyto tv array int arrayindex defer error checking to array copy array copy dict values array arrayindex dict count void icollection copyto array array int index if array null array rank throw new argumentexception sr arg rankmultidimnotsupported nameof array try defer error checking to array copy array copy dict values array index dict count catch arraytypemismatchexception throw new argumentexception sr argument invalidarraytype nameof array public void insert int index tv value throw new notsupportedexception sr notsupported sortedlistnestedwrite public tv this get return dict getbyindex index set throw new notsupportedexception sr notsupported sortedlistnestedwrite public ienumerator getenumerator return new sortedlistvalueenumerator dict ienumerator ienumerable getenumerator return new sortedlistvalueenumerator dict public int indexof tv value return array indexof dict values value dict count public bool remove tv value throw new notsupportedexception sr notsupported sortedlistnestedwrite return false public void removeat int index throw new notsupportedexception sr notsupported sortedlistnestedwrite public static void main var sortedlist new sortedlist var msg hello world sortedlist count console writeline msg try other c samples from the button in the top right expected result no runtime error actual result system exception typeerror cannot read property interfaces of undefined at object createinheritors at function i staticinit at object init at object bridge assembly at | 1 |
51,332 | 13,207,435,860 | IssuesEvent | 2020-08-14 23:05:34 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | Time goes backwards: fatal lock of Server. (Trac #233) | Incomplete Migration Migrated from Trac defect jeb + pnf | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/233">https://code.icecube.wisc.edu/projects/icecube/ticket/233</a>, reported by blaufussand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:33:31",
"_ts": "1351704811000000",
"description": "Seems a perhaps overly agressive time continuity check when measuring delays of clients and such can occasionally yield:\n\n\n2010-11-19 04:59:43 [GMT] FATAL ServerQueue : /scratch/blaufuss/jeb/V10-09-00/src/jebserver/private/jebserver/ServerQueue.cxx:640 oups - time goes backwards!?!\n\n\nThis is likely not real (no indication of a ntpd update at that time)\n\nInvestigate, sanitize the error condition and maybe keep processing?\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "worksforme",
"time": "2010-12-01T21:17:38",
"component": "jeb + pnf",
"summary": "Time goes backwards: fatal lock of Server.",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Time goes backwards: fatal lock of Server. (Trac #233) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/233">https://code.icecube.wisc.edu/projects/icecube/ticket/233</a>, reported by blaufussand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:33:31",
"_ts": "1351704811000000",
"description": "Seems a perhaps overly agressive time continuity check when measuring delays of clients and such can occasionally yield:\n\n\n2010-11-19 04:59:43 [GMT] FATAL ServerQueue : /scratch/blaufuss/jeb/V10-09-00/src/jebserver/private/jebserver/ServerQueue.cxx:640 oups - time goes backwards!?!\n\n\nThis is likely not real (no indication of a ntpd update at that time)\n\nInvestigate, sanitize the error condition and maybe keep processing?\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "worksforme",
"time": "2010-12-01T21:17:38",
"component": "jeb + pnf",
"summary": "Time goes backwards: fatal lock of Server.",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
| defect | time goes backwards fatal lock of server trac migrated from json status closed changetime ts description seems a perhaps overly agressive time continuity check when measuring delays of clients and such can occasionally yield n n fatal serverqueue scratch blaufuss jeb src jebserver private jebserver serverqueue cxx oups time goes backwards n n nthis is likely not real no indication of a ntpd update at that time n ninvestigate sanitize the error condition and maybe keep processing n n reporter blaufuss cc resolution worksforme time component jeb pnf summary time goes backwards fatal lock of server priority normal keywords milestone owner tschmidt type defect | 1 |
11,342 | 2,649,274,247 | IssuesEvent | 2015-03-14 19:08:39 | perfsonar/perl-perfsonar | https://api.github.com/repos/perfsonar/perl-perfsonar | closed | LSCacheDaemon does not work through web proxies | Milestone-Release3.4.2 Priority-Medium Type-Defect | Original [issue 966](https://code.google.com/p/perfsonar-ps/issues/detail?id=966) created by arlake228 on 2014-08-04T12:23:20.000Z:
Hello,
Our site employs a site-wide HTTP proxy. We are using PerfSONAR-pS to monitor performance at various points inside the site, and because of the mandatory proxy, none of our nodes can retrieve the PS Community host lists.
(note that we do not run tests through the web proxy - this is for the community cache!)
I have attached a set of small patches that fix the problem. The fix employs a global configuration file in /etc/sysconfig and a few changes to LSCacheDaemon to get the HTTP_PROXY environment variable picked up.
I would have submitted these directly via git but I had trouble figuring out how you should package the two new files; /etc/sysconfig/proxy and /etc/profile.d/proxy.sh
<b>What steps will reproduce the problem?</b>
1. Put a PerfSONAR-pS instance on a network that does not allow port 80/443 traffic
2. Make a HTTP proxy available on the network
3. There is no configuration in PerfSONAR-pS that will let you configure a HTTP proxy for the LSCacheDaemon
<b>What version of the product are you using? On what operating system?</b>
> Internet2 pS-Performance Toolkit v3.3.2 netinstall
Thanks,
Alex White
Diamond Light Source
Diamond.ac.uk | 1.0 | LSCacheDaemon does not work through web proxies - Original [issue 966](https://code.google.com/p/perfsonar-ps/issues/detail?id=966) created by arlake228 on 2014-08-04T12:23:20.000Z:
Hello,
Our site employs a site-wide HTTP proxy. We are using PerfSONAR-pS to monitor performance at various points inside the site, and because of the mandatory proxy, none of our nodes can retrieve the PS Community host lists.
(note that we do not run tests through the web proxy - this is for the community cache!)
I have attached a set of small patches that fix the problem. The fix employs a global configuration file in /etc/sysconfig and a few changes to LSCacheDaemon to get the HTTP_PROXY environment variable picked up.
I would have submitted these directly via git but I had trouble figuring out how you should package the two new files; /etc/sysconfig/proxy and /etc/profile.d/proxy.sh
<b>What steps will reproduce the problem?</b>
1. Put a PerfSONAR-pS instance on a network that does not allow port 80/443 traffic
2. Make a HTTP proxy available on the network
3. There is no configuration in PerfSONAR-pS that will let you configure a HTTP proxy for the LSCacheDaemon
<b>What version of the product are you using? On what operating system?</b>
> Internet2 pS-Performance Toolkit v3.3.2 netinstall
Thanks,
Alex White
Diamond Light Source
Diamond.ac.uk | defect | lscachedaemon does not work through web proxies original created by on hello our site employs a site wide http proxy we are using perfsonar ps to monitor performance at various points inside the site and because of the mandatory proxy none of our nodes can retrieve the ps community host lists note that we do not run tests through the web proxy this is for the community cache i have attached a set of small patches that fix the problem the fix employs a global configuration file in etc sysconfig and a few changes to lscachedaemon to get the http proxy environment variable picked up i would have submitted these directly via git but i had trouble figuring out how you should package the two new files etc sysconfig proxy and etc profile d proxy sh what steps will reproduce the problem put a perfsonar ps instance on a network that does not allow port traffic make a http proxy available on the network there is no configuration in perfsonar ps that will let you configure a http proxy for the lscachedaemon what version of the product are you using on what operating system gt ps performance toolkit netinstall thanks alex white diamond light source diamond ac uk | 1 |
64,892 | 18,954,038,716 | IssuesEvent | 2021-11-18 18:02:46 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | BUG: openblas 0.3.9 shipped with the scipy 1.7.1 wheel can cause segfaults when it fails to detect the correct CPU architecture (Prescott instead o Haswell) | defect | ### Describe your issue.
Recent Intel CPUs detected as `SkylakeX` by recent versions of OpenBLAS are wrongfully detected as `Prescott` by OpenBLAS 0.3.9 shipped with scipy 1.7.1.
This in turns can cause segfaults when calling the `ddot_k_PRESCOTT` routine on a readonly memory buffer allocated by `numpy.memmap` with `mode="r"`.
EDIT: the more probable cause is that the data is not memory aligned in this test case.
See: https://github.com/scikit-learn/scikit-learn/issues/21361 for details an reproducing code using the scikit-learn test suite.
### Error message
```shell
(gdb) bt
#0 0x00007fffdbe31c31 in ddot_k_PRESCOTT () from /home/aperez/dev/sandbox/test/scikit-learn/venv/lib/python3.8/site-packages/scipy/spatial/../../scipy.libs/libopenblasp-r0-085ca80a.3.9.so
#1 0x00007fffc893b39a in ?? () from /home/aperez/dev/sandbox/test/scikit-learn/venv/lib/python3.8/site-packages/scipy/linalg/cython_blas.cpython-38-x86_64-linux-gnu.so
#2 0x00007fffc89226e0 in ?? () from /home/aperez/dev/sandbox/test/scikit-learn/venv/lib/python3.8/site-packages/scipy/linalg/cython_blas.cpython-38-x86_64-linux-gnu.so
#3 0x00007fffbe547aca in __pyx_fuse_1__pyx_f_7sklearn_5utils_12_cython_blas__dot (__pyx_v_n=<optimized out>, __pyx_v_x=<optimized out>, __pyx_v_incx=<optimized out>, __pyx_v_y=<optimized out>,
__pyx_v_incy=<optimized out>) at sklearn/utils/_cython_blas.c:2861
#4 0x00007fffbe258327 in svm::Kernel::Kernel (this=0x7fffffff6030, l=<optimized out>, x_=<optimized out>, param=..., blas_functions=0x7fffffff64e0) at sklearn/svm/src/libsvm/svm.cpp:394
#5 0x00007fffbe25b394 in svm::SVC_Q::SVC_Q (blas_functions=0x7fffffff64e0, y_=0x555556ff0fd0 '\001' <repeats 100 times>, '\377' <repeats 100 times>..., param=..., prob=..., this=0x7fffffff6030)
at sklearn/svm/src/libsvm/svm.cpp:1682
#6 svm::solve_nu_svc (blas_functions=0x7fffffff64e0, si=0x7fffffff6000, alpha=0x555557450160, param=0x7fffffff69a0, prob=0x7fffffff6360) at sklearn/svm/src/libsvm/svm.cpp:1682
#7 svm::svm_train_one (prob=0x7fffffff6360, param=0x7fffffff69a0, Cp=<optimized out>, Cn=<optimized out>, status=0x7fffffff64d4, blas_functions=0x7fffffff64e0)
at sklearn/svm/src/libsvm/svm.cpp:1856
#8 0x00007fffbe264cde in svm_train (prob=0x7fffffff6340, prob@entry=0x7fffffff6500, param=param@entry=0x7fffffff69a0, status=status@entry=0x7fffffff64d4,
blas_functions=blas_functions@entry=0x7fffffff64e0) at sklearn/svm/src/libsvm/svm.cpp:2504
#9 0x00007fffbe23e921 in __pyx_pf_7sklearn_3svm_7_libsvm_fit (__pyx_v_X=__pyx_v_X@entry=0x7fffbb041db0, __pyx_v_Y=__pyx_v_Y@entry=0x7fffbb041c90, __pyx_v_svm_type=__pyx_v_svm_type@entry=1,
__pyx_v_kernel=__pyx_v_kernel@entry=0x7fffc3052730, __pyx_v_degree=__pyx_v_degree@entry=3, __pyx_v_gamma=__pyx_v_gamma@entry=0.53188777537536391, __pyx_v_coef0=__pyx_v_coef0@entry=0,
__pyx_v_tol=__pyx_v_tol@entry=0.001, __pyx_v_C=__pyx_v_C@entry=0, __pyx_v_nu=0.5, __pyx_v_epsilon=0, __pyx_v_class_weight=__pyx_v_class_weight@entry=0x7fffbb041f30,
__pyx_v_sample_weight=0x7fffbb041e10, __pyx_v_shrinking=1, __pyx_v_probability=0, __pyx_v_cache_size=200, __pyx_v_max_iter=-1, __pyx_v_random_seed=209652396, __pyx_self=<optimized out>)
[...]
```
### SciPy/NumPy/Python version information
1.7.1 1.21.2 3.8.2 | 1.0 | BUG: openblas 0.3.9 shipped with the scipy 1.7.1 wheel can cause segfaults when it fails to detect the correct CPU architecture (Prescott instead o Haswell) - ### Describe your issue.
Recent Intel CPUs detected as `SkylakeX` by recent versions of OpenBLAS are wrongfully detected as `Prescott` by OpenBLAS 0.3.9 shipped with scipy 1.7.1.
This in turns can cause segfaults when calling the `ddot_k_PRESCOTT` routine on a readonly memory buffer allocated by `numpy.memmap` with `mode="r"`.
EDIT: the more probable cause is that the data is not memory aligned in this test case.
See: https://github.com/scikit-learn/scikit-learn/issues/21361 for details an reproducing code using the scikit-learn test suite.
### Error message
```shell
(gdb) bt
#0 0x00007fffdbe31c31 in ddot_k_PRESCOTT () from /home/aperez/dev/sandbox/test/scikit-learn/venv/lib/python3.8/site-packages/scipy/spatial/../../scipy.libs/libopenblasp-r0-085ca80a.3.9.so
#1 0x00007fffc893b39a in ?? () from /home/aperez/dev/sandbox/test/scikit-learn/venv/lib/python3.8/site-packages/scipy/linalg/cython_blas.cpython-38-x86_64-linux-gnu.so
#2 0x00007fffc89226e0 in ?? () from /home/aperez/dev/sandbox/test/scikit-learn/venv/lib/python3.8/site-packages/scipy/linalg/cython_blas.cpython-38-x86_64-linux-gnu.so
#3 0x00007fffbe547aca in __pyx_fuse_1__pyx_f_7sklearn_5utils_12_cython_blas__dot (__pyx_v_n=<optimized out>, __pyx_v_x=<optimized out>, __pyx_v_incx=<optimized out>, __pyx_v_y=<optimized out>,
__pyx_v_incy=<optimized out>) at sklearn/utils/_cython_blas.c:2861
#4 0x00007fffbe258327 in svm::Kernel::Kernel (this=0x7fffffff6030, l=<optimized out>, x_=<optimized out>, param=..., blas_functions=0x7fffffff64e0) at sklearn/svm/src/libsvm/svm.cpp:394
#5 0x00007fffbe25b394 in svm::SVC_Q::SVC_Q (blas_functions=0x7fffffff64e0, y_=0x555556ff0fd0 '\001' <repeats 100 times>, '\377' <repeats 100 times>..., param=..., prob=..., this=0x7fffffff6030)
at sklearn/svm/src/libsvm/svm.cpp:1682
#6 svm::solve_nu_svc (blas_functions=0x7fffffff64e0, si=0x7fffffff6000, alpha=0x555557450160, param=0x7fffffff69a0, prob=0x7fffffff6360) at sklearn/svm/src/libsvm/svm.cpp:1682
#7 svm::svm_train_one (prob=0x7fffffff6360, param=0x7fffffff69a0, Cp=<optimized out>, Cn=<optimized out>, status=0x7fffffff64d4, blas_functions=0x7fffffff64e0)
at sklearn/svm/src/libsvm/svm.cpp:1856
#8 0x00007fffbe264cde in svm_train (prob=0x7fffffff6340, prob@entry=0x7fffffff6500, param=param@entry=0x7fffffff69a0, status=status@entry=0x7fffffff64d4,
blas_functions=blas_functions@entry=0x7fffffff64e0) at sklearn/svm/src/libsvm/svm.cpp:2504
#9 0x00007fffbe23e921 in __pyx_pf_7sklearn_3svm_7_libsvm_fit (__pyx_v_X=__pyx_v_X@entry=0x7fffbb041db0, __pyx_v_Y=__pyx_v_Y@entry=0x7fffbb041c90, __pyx_v_svm_type=__pyx_v_svm_type@entry=1,
__pyx_v_kernel=__pyx_v_kernel@entry=0x7fffc3052730, __pyx_v_degree=__pyx_v_degree@entry=3, __pyx_v_gamma=__pyx_v_gamma@entry=0.53188777537536391, __pyx_v_coef0=__pyx_v_coef0@entry=0,
__pyx_v_tol=__pyx_v_tol@entry=0.001, __pyx_v_C=__pyx_v_C@entry=0, __pyx_v_nu=0.5, __pyx_v_epsilon=0, __pyx_v_class_weight=__pyx_v_class_weight@entry=0x7fffbb041f30,
__pyx_v_sample_weight=0x7fffbb041e10, __pyx_v_shrinking=1, __pyx_v_probability=0, __pyx_v_cache_size=200, __pyx_v_max_iter=-1, __pyx_v_random_seed=209652396, __pyx_self=<optimized out>)
[...]
```
### SciPy/NumPy/Python version information
1.7.1 1.21.2 3.8.2 | defect | bug openblas shipped with the scipy wheel can cause segfaults when it fails to detect the correct cpu architecture prescott instead o haswell describe your issue recent intel cpus detected as skylakex by recent versions of openblas are wrongfully detected as prescott by openblas shipped with scipy this in turns can cause segfaults when calling the ddot k prescott routine on a readonly memory buffer allocated by numpy memmap with mode r edit the more probable cause is that the data is not memory aligned in this test case see for details an reproducing code using the scikit learn test suite error message shell gdb bt in ddot k prescott from home aperez dev sandbox test scikit learn venv lib site packages scipy spatial scipy libs libopenblasp so in from home aperez dev sandbox test scikit learn venv lib site packages scipy linalg cython blas cpython linux gnu so in from home aperez dev sandbox test scikit learn venv lib site packages scipy linalg cython blas cpython linux gnu so in pyx fuse pyx f cython blas dot pyx v n pyx v x pyx v incx pyx v y pyx v incy at sklearn utils cython blas c in svm kernel kernel this l x param blas functions at sklearn svm src libsvm svm cpp in svm svc q svc q blas functions y param prob this at sklearn svm src libsvm svm cpp svm solve nu svc blas functions si alpha param prob at sklearn svm src libsvm svm cpp svm svm train one prob param cp cn status blas functions at sklearn svm src libsvm svm cpp in svm train prob prob entry param param entry status status entry blas functions blas functions entry at sklearn svm src libsvm svm cpp in pyx pf libsvm fit pyx v x pyx v x entry pyx v y pyx v y entry pyx v svm type pyx v svm type entry pyx v kernel pyx v kernel entry pyx v degree pyx v degree entry pyx v gamma pyx v gamma entry pyx v pyx v entry pyx v tol pyx v tol entry pyx v c pyx v c entry pyx v nu pyx v epsilon pyx v class weight pyx v class weight entry pyx v sample weight pyx v shrinking pyx v probability pyx v cache size pyx v max iter pyx v random seed pyx self scipy numpy python version information | 1 |
431,581 | 30,241,609,168 | IssuesEvent | 2023-07-06 13:52:05 | eu-cdse/documentation | https://api.github.com/repos/eu-cdse/documentation | closed | Improve link color | documentation | Currently, link colors have a relatively light green color (`#77cc09`). In some contexts, like on a gray background, this results in a pretty low contrast text, hurting readability.
e.g.:

Interestingly, I noticed that the readability in this example is considerably worse on my external monitor than on my laptop's display. | 1.0 | Improve link color - Currently, link colors have a relatively light green color (`#77cc09`). In some contexts, like on a gray background, this results in a pretty low contrast text, hurting readability.
e.g.:

Interestingly, I noticed that the readability in this example is considerably worse on my external monitor than on my laptop's display. | non_defect | improve link color currently link colors have a relatively light green color in some contexts like on a gray background this results in a pretty low contrast text hurting readability e g interestingly i noticed that the readability in this example is considerably worse on my external monitor than on my laptop s display | 0 |
40,258 | 9,937,262,239 | IssuesEvent | 2019-07-02 21:22:48 | jccastillo0007/eFacturaT | https://api.github.com/repos/jccastillo0007/eFacturaT | opened | Condominios-CxP-Reporte general, no incluye las facturas o egresos ya pagados | bug defect | Solo incluye los documentos de egreso o facturas de proveedores que no han sido pagados.
Recordar que ahí se deben sumarizar todos los documentos de egreso y facturas , por proveedor.... y eso se muestra en la columna abono, hayan sido pagadas o no.
La sumatoria de pagos realizados por proveedor, se deben mostrar en la columna cargo. | 1.0 | Condominios-CxP-Reporte general, no incluye las facturas o egresos ya pagados - Solo incluye los documentos de egreso o facturas de proveedores que no han sido pagados.
Recordar que ahí se deben sumarizar todos los documentos de egreso y facturas , por proveedor.... y eso se muestra en la columna abono, hayan sido pagadas o no.
La sumatoria de pagos realizados por proveedor, se deben mostrar en la columna cargo. | defect | condominios cxp reporte general no incluye las facturas o egresos ya pagados solo incluye los documentos de egreso o facturas de proveedores que no han sido pagados recordar que ahí se deben sumarizar todos los documentos de egreso y facturas por proveedor y eso se muestra en la columna abono hayan sido pagadas o no la sumatoria de pagos realizados por proveedor se deben mostrar en la columna cargo | 1 |
61,599 | 17,023,735,307 | IssuesEvent | 2021-07-03 03:33:47 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | yevaud: tile server down? "more osm coming soon" | Component: website Priority: blocker Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 8.18am, Monday, 25th July 2011]**
Currently the tile server seems to be down:
http://wiki.openstreetmap.org/wiki/Platform_Status
Users experience lots of "more osm coming soon" tiles being "delivered".
The inode table use seems a bit off on yevaud
http://munin.openstreetmap.org/openstreetmap/yevaud.openstreetmap/open_inodes.html
| 1.0 | yevaud: tile server down? "more osm coming soon" - **[Submitted to the original trac issue database at 8.18am, Monday, 25th July 2011]**
Currently the tile server seems to be down:
http://wiki.openstreetmap.org/wiki/Platform_Status
Users experience lots of "more osm coming soon" tiles being "delivered".
The inode table use seems a bit off on yevaud
http://munin.openstreetmap.org/openstreetmap/yevaud.openstreetmap/open_inodes.html
| defect | yevaud tile server down more osm coming soon currently the tile server seems to be down users experience lots of more osm coming soon tiles being delivered the inode table use seems a bit off on yevaud | 1 |
53,046 | 13,260,845,681 | IssuesEvent | 2020-08-20 18:51:33 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | can't find python libs > python 2.6.x (Trac #627) | Migrated from Trac cmake defect | cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)
currently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).
this '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.
probably be able to solve this w/ i3-tools-v4 (#278)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/627">https://code.icecube.wisc.edu/projects/icecube/ticket/627</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-05-11T19:50:31",
"_ts": "1305143431000000",
"description": "cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)\n\ncurrently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).\n\nthis '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.\n\nprobably be able to solve this w/ i3-tools-v4 (#278)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2011-04-30T16:22:01",
"component": "cmake",
"summary": "can't find python libs > python 2.6.x",
"priority": "major",
"keywords": "cmake python freebsd rhel",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | can't find python libs > python 2.6.x (Trac #627) - cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)
currently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).
this '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.
probably be able to solve this w/ i3-tools-v4 (#278)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/627">https://code.icecube.wisc.edu/projects/icecube/ticket/627</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-05-11T19:50:31",
"_ts": "1305143431000000",
"description": "cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)\n\ncurrently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).\n\nthis '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.\n\nprobably be able to solve this w/ i3-tools-v4 (#278)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2011-04-30T16:22:01",
"component": "cmake",
"summary": "can't find python libs > python 2.6.x",
"priority": "major",
"keywords": "cmake python freebsd rhel",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | can t find python libs python x trac cmake x only has support for python x in its pythonfindx cmake modules hard coded values this seems to have been fixed in cmake x but is currently untested by icecube i personally use cmake x but not python x currently i solve this on the build bots by hard coding python paths in a cite cmake dir file on akuma external python and beastie freebsd freebsd port install of python in usr local this can not be solved by including the cmake findpythonx modules as we d get into dependency hell w the cmake modules probably be able to solve this w tools migrated from json status closed changetime ts description cmake x only has support for python x in its pythonfindx cmake modules hard coded values this seems to have been fixed in cmake x but is currently untested by icecube i personally use cmake x but not python x n ncurrently i solve this on the build bots by hard coding python paths in a cite cmake dir file on akuma external python and beastie freebsd freebsd port install of python in usr local n nthis can not be solved by including the cmake findpythonx modules as we d get into dependency hell w the cmake modules n nprobably be able to solve this w tools reporter nega cc resolution fixed time component cmake summary can t find python libs python x priority major keywords cmake python freebsd rhel milestone owner nega type defect | 1 |
125,715 | 4,963,955,622 | IssuesEvent | 2016-12-03 14:35:16 | chartjs/Chart.js | https://api.github.com/repos/chartjs/Chart.js | closed | Option to make legend not visible if value is 0 | Category: Enhancement Help wanted Priority: p2 | Is there a way for an option in the legend to make the label hidden if the value is not there or the value is 0?
| 1.0 | Option to make legend not visible if value is 0 - Is there a way for an option in the legend to make the label hidden if the value is not there or the value is 0?
| non_defect | option to make legend not visible if value is is there a way for an option in the legend to make the label hidden if the value is not there or the value is | 0 |
188,536 | 14,446,673,770 | IssuesEvent | 2020-12-08 01:51:46 | doctrine/orm | https://api.github.com/repos/doctrine/orm | closed | Wrong insert SQL command order | Bug Missing Tests | The schema has following structure:
User `<--Join Inheritance--` Store `<--OneToMany/ManyToOne--` Stock ` --ManyToOne/undirectional-->` Address.
Right insert order is: User, Store, Address, Stock
But doctrine tries to insert objects in this order: Address, Stock, etc. In Stock insertion SQL command fails because `store_id` is NOT NULL field.
It seems like it doesn't take into account that Stock's `store_id` field has NOT NULL flag so Store should be created before Stock. | 1.0 | Wrong insert SQL command order - The schema has following structure:
User `<--Join Inheritance--` Store `<--OneToMany/ManyToOne--` Stock ` --ManyToOne/undirectional-->` Address.
Right insert order is: User, Store, Address, Stock
But doctrine tries to insert objects in this order: Address, Stock, etc. In Stock insertion SQL command fails because `store_id` is NOT NULL field.
It seems like it doesn't take into account that Stock's `store_id` field has NOT NULL flag so Store should be created before Stock. | non_defect | wrong insert sql command order the schema has following structure user address right insert order is user store address stock but doctrine tries to insert objects in this order address stock etc in stock insertion sql command fails because store id is not null field it seems like it doesn t take into account that stock s store id field has not null flag so store should be created before stock | 0 |
49,848 | 13,187,280,826 | IssuesEvent | 2020-08-13 02:54:59 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | DOMLauncher -fext-numeric-literals breaks compilation with clang (Trac #2176) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2176">https://code.icecube.wisc.edu/ticket/2176</a>, reported by tianlu and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-07-20T09:41:47",
"description": "Changeset 163998 [http://code.icecube.wisc.edu/projects/icecube/changeset/163998/IceCube] doesn't play nicely with clang. I get the following error when trying to build on a Mac.\n{{{\nclang: error: unknown argument: '-fext-numeric-literals' \nmake[2]: *** [DOMLauncher/CMakeFiles/DOMLauncher.dir/private/DOMLauncher/DOMLauncher.cxx.o] Error 1\nmake[1]: *** [DOMLauncher/CMakeFiles/DOMLauncher.dir/all] Error 2\nmake[1]: *** Waiting for unfinished jobs.... \n}}}\n",
"reporter": "tianlu",
"cc": "chraab",
"resolution": "fixed",
"_ts": "1532079707579233",
"component": "cmake",
"summary": "DOMLauncher -fext-numeric-literals breaks compilation with clang",
"priority": "minor",
"keywords": "",
"time": "2018-07-20T02:06:35",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | DOMLauncher -fext-numeric-literals breaks compilation with clang (Trac #2176) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2176">https://code.icecube.wisc.edu/ticket/2176</a>, reported by tianlu and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-07-20T09:41:47",
"description": "Changeset 163998 [http://code.icecube.wisc.edu/projects/icecube/changeset/163998/IceCube] doesn't play nicely with clang. I get the following error when trying to build on a Mac.\n{{{\nclang: error: unknown argument: '-fext-numeric-literals' \nmake[2]: *** [DOMLauncher/CMakeFiles/DOMLauncher.dir/private/DOMLauncher/DOMLauncher.cxx.o] Error 1\nmake[1]: *** [DOMLauncher/CMakeFiles/DOMLauncher.dir/all] Error 2\nmake[1]: *** Waiting for unfinished jobs.... \n}}}\n",
"reporter": "tianlu",
"cc": "chraab",
"resolution": "fixed",
"_ts": "1532079707579233",
"component": "cmake",
"summary": "DOMLauncher -fext-numeric-literals breaks compilation with clang",
"priority": "minor",
"keywords": "",
"time": "2018-07-20T02:06:35",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | domlauncher fext numeric literals breaks compilation with clang trac migrated from json status closed changetime description changeset doesn t play nicely with clang i get the following error when trying to build on a mac n nclang error unknown argument fext numeric literals nmake error nmake error nmake waiting for unfinished jobs n n reporter tianlu cc chraab resolution fixed ts component cmake summary domlauncher fext numeric literals breaks compilation with clang priority minor keywords time milestone owner type defect | 1 |
641,829 | 20,841,700,267 | IssuesEvent | 2022-03-21 01:20:40 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | opened | IX: when toggling filters go back to page 1 | Bug :lady_beetle: Priority: Medium Sprint-Ready | If the user is in a page other than 1, using the filters can toss less results than the current pagination setup, leaving the user with a blank table.
IMO a quick solution would be to send the user back to page one after filtering. | 1.0 | IX: when toggling filters go back to page 1 - If the user is in a page other than 1, using the filters can toss less results than the current pagination setup, leaving the user with a blank table.
IMO a quick solution would be to send the user back to page one after filtering. | non_defect | ix when toggling filters go back to page if the user is in a page other than using the filters can toss less results than the current pagination setup leaving the user with a blank table imo a quick solution would be to send the user back to page one after filtering | 0 |
5,772 | 2,610,215,113 | IssuesEvent | 2015-02-26 19:08:37 | chrsmith/somefinders | https://api.github.com/repos/chrsmith/somefinders | opened | инструкцию к хлебопечке bifinett kh 1171.rar | auto-migrated Priority-Medium Type-Defect | ```
'''Гертруд Фомин'''
Привет всем не подскажите где можно найти
.инструкцию к хлебопечке bifinett kh 1171.rar. как то
выкладывали уже
'''Гарибальди Марков'''
Вот хороший сайт где можно скачать
http://bit.ly/1aWUWyM
'''Владелин Доронин'''
Спасибо вроде то но просит телефон вводить
'''Аксён Шаров'''
Не это не влияет на баланс
'''Володар Большаков'''
Неа все ок у меня ничего не списало
Информация о файле: инструкцию к
хлебопечке bifinett kh 1171.rar
Загружен: В этом месяце
Скачан раз: 1369
Рейтинг: 768
Средняя скорость скачивания: 515
Похожих файлов: 34
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:14 | 1.0 | инструкцию к хлебопечке bifinett kh 1171.rar - ```
'''Гертруд Фомин'''
Привет всем не подскажите где можно найти
.инструкцию к хлебопечке bifinett kh 1171.rar. как то
выкладывали уже
'''Гарибальди Марков'''
Вот хороший сайт где можно скачать
http://bit.ly/1aWUWyM
'''Владелин Доронин'''
Спасибо вроде то но просит телефон вводить
'''Аксён Шаров'''
Не это не влияет на баланс
'''Володар Большаков'''
Неа все ок у меня ничего не списало
Информация о файле: инструкцию к
хлебопечке bifinett kh 1171.rar
Загружен: В этом месяце
Скачан раз: 1369
Рейтинг: 768
Средняя скорость скачивания: 515
Похожих файлов: 34
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:14 | defect | инструкцию к хлебопечке bifinett kh rar гертруд фомин привет всем не подскажите где можно найти инструкцию к хлебопечке bifinett kh rar как то выкладывали уже гарибальди марков вот хороший сайт где можно скачать владелин доронин спасибо вроде то но просит телефон вводить аксён шаров не это не влияет на баланс володар большаков неа все ок у меня ничего не списало информация о файле инструкцию к хлебопечке bifinett kh rar загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at | 1 |
40,357 | 9,966,992,610 | IssuesEvent | 2019-07-08 12:38:14 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | p-multiselect disabled and selected option gets un-selected by the select all checkbox in header | defect | **I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
https://github-nanxuo.stackblitz.io
**Current behavior**
If a SelectItem is disabled and also selected by default, if you choose the select all checkbox in the p-multiSelect element header it will un-select the disabled SelectItem and it cannot be re-selected.
**Expected behavior**
Any disabled selectItems should not be affected by the select all checkbox.
**Minimal reproduction of the problem with instructions**
Set a SelectItem to disabled.
```
this.cars = [
{label: 'Audi', value: 'Audi'},
{label: 'BMW', value: 'BMW'},
{label: 'Fiat', value: 'Fiat', disabled: true},
{label: 'Ford', value: 'Ford'},
{label: 'Honda', value: 'Honda'},
{label: 'Jaguar', value: 'Jaguar'},
{label: 'Mercedes', value: 'Mercedes', disabled: true},
{label: 'Renault', value: 'Renault'},
{label: 'VW', value: 'VW'},
{label: 'Volvo', value: 'Volvo'},
];
```
Set that same SelectItem as selected.
```
selectedCars1: string[] = [
'Audi',
'BMW',
'Fiat'
];
```
Check the select all checkbox in the multiselect header and see that the disabled SelectItem has been un-selected and cannot be re-selected.
**What is the motivation / use case for changing the behavior?**
In certain instances an interface may need an item that is displayed as checked but should not be changed. For example, an interface where assigning users to roles, but all users have a default role that should not be changed by anyone..
**Please tell us about your environment:**
MacOS, Visual Studio Code, npm
* **Angular version:** 7.2.8
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 7.1.0
This can be replicated on the master branch
* **Browser:** Chrome 73.0.3683.86
* **Language:** TypeScript 3.2.4 | ES5
| 1.0 | p-multiselect disabled and selected option gets un-selected by the select all checkbox in header - **I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
https://github-nanxuo.stackblitz.io
**Current behavior**
If a SelectItem is disabled and also selected by default, if you choose the select all checkbox in the p-multiSelect element header it will un-select the disabled SelectItem and it cannot be re-selected.
**Expected behavior**
Any disabled selectItems should not be affected by the select all checkbox.
**Minimal reproduction of the problem with instructions**
Set a SelectItem to disabled.
```
this.cars = [
{label: 'Audi', value: 'Audi'},
{label: 'BMW', value: 'BMW'},
{label: 'Fiat', value: 'Fiat', disabled: true},
{label: 'Ford', value: 'Ford'},
{label: 'Honda', value: 'Honda'},
{label: 'Jaguar', value: 'Jaguar'},
{label: 'Mercedes', value: 'Mercedes', disabled: true},
{label: 'Renault', value: 'Renault'},
{label: 'VW', value: 'VW'},
{label: 'Volvo', value: 'Volvo'},
];
```
Set that same SelectItem as selected.
```
selectedCars1: string[] = [
'Audi',
'BMW',
'Fiat'
];
```
Check the select all checkbox in the multiselect header and see that the disabled SelectItem has been un-selected and cannot be re-selected.
**What is the motivation / use case for changing the behavior?**
In certain instances an interface may need an item that is displayed as checked but should not be changed. For example, an interface where assigning users to roles, but all users have a default role that should not be changed by anyone..
**Please tell us about your environment:**
MacOS, Visual Studio Code, npm
* **Angular version:** 7.2.8
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 7.1.0
This can be replicated on the master branch
* **Browser:** Chrome 73.0.3683.86
* **Language:** TypeScript 3.2.4 | ES5
| defect | p multiselect disabled and selected option gets un selected by the select all checkbox in header i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports current behavior if a selectitem is disabled and also selected by default if you choose the select all checkbox in the p multiselect element header it will un select the disabled selectitem and it cannot be re selected expected behavior any disabled selectitems should not be affected by the select all checkbox minimal reproduction of the problem with instructions set a selectitem to disabled this cars label audi value audi label bmw value bmw label fiat value fiat disabled true label ford value ford label honda value honda label jaguar value jaguar label mercedes value mercedes disabled true label renault value renault label vw value vw label volvo value volvo set that same selectitem as selected string audi bmw fiat check the select all checkbox in the multiselect header and see that the disabled selectitem has been un selected and cannot be re selected what is the motivation use case for changing the behavior in certain instances an interface may need an item that is displayed as checked but should not be changed for example an interface where assigning users to roles but all users have a default role that should not be changed by anyone please tell us about your environment macos visual studio code npm angular version primeng version this can be replicated on the master branch browser chrome language typescript | 1 |
133,796 | 18,953,984,123 | IssuesEvent | 2021-11-18 17:59:03 | CDCgov/prime-simplereport | https://api.github.com/repos/CDCgov/prime-simplereport | opened | Adjust duplicate check endpoint to be First Name, Last Name, and Date of Birth | Design Better Data | ### WHY?
The current duplicate endpoint checks if a user is a duplicate based on their **first name, last name, DOB, and Zipcode**. Checking for zipcode creates more possibilities of missing duplicates due to:
- varying zipcode lengths (e.g. 94606 vs. 94606-2354 vs. 946062354)
- some users are highly transient, which means frequent changes to their Zipcode and the higher likelihood of users not receiving our duplicate prevention patterns.
By removing the zipcode endpoint, users are more likely to catch duplicate registrations sooner as they won't have to fill out the "contact information" section in order for our duplicate check to work.
**AS A** SR user
**I WANT** to be notified when I may be registering a duplicate patient
**SO THAT** I can stop the registration and in turn prevent duplicate patient profiles in my facility
---
**WHEN** the user enters the first name, last name, and date of birth in the "add new person"/self-registration workflow
**THEN** run a duplicate check
| 1.0 | Adjust duplicate check endpoint to be First Name, Last Name, and Date of Birth - ### WHY?
The current duplicate endpoint checks if a user is a duplicate based on their **first name, last name, DOB, and Zipcode**. Checking for zipcode creates more possibilities of missing duplicates due to:
- varying zipcode lengths (e.g. 94606 vs. 94606-2354 vs. 946062354)
- some users are highly transient, which means frequent changes to their Zipcode and the higher likelihood of users not receiving our duplicate prevention patterns.
By removing the zipcode endpoint, users are more likely to catch duplicate registrations sooner as they won't have to fill out the "contact information" section in order for our duplicate check to work.
**AS A** SR user
**I WANT** to be notified when I may be registering a duplicate patient
**SO THAT** I can stop the registration and in turn prevent duplicate patient profiles in my facility
---
**WHEN** the user enters the first name, last name, and date of birth in the "add new person"/self-registration workflow
**THEN** run a duplicate check
| non_defect | adjust duplicate check endpoint to be first name last name and date of birth why the current duplicate endpoint checks if a user is a duplicate based on their first name last name dob and zipcode checking for zipcode creates more possibilities of missing duplicates due to varying zipcode lengths e g vs vs some users are highly transient which means frequent changes to their zipcode and the higher likelihood of users not receiving our duplicate prevention patterns by removing the zipcode endpoint users are more likely to catch duplicate registrations sooner as they won t have to fill out the contact information section in order for our duplicate check to work as a sr user i want to be notified when i may be registering a duplicate patient so that i can stop the registration and in turn prevent duplicate patient profiles in my facility when the user enters the first name last name and date of birth in the add new person self registration workflow then run a duplicate check | 0 |
438,640 | 30,653,620,842 | IssuesEvent | 2023-07-25 10:32:32 | pulgamecanica/42Portfolio | https://api.github.com/repos/pulgamecanica/42Portfolio | closed | Wiki has a minor typo | documentation | I tried to create a PR, but I don't have access permissions to the wiki repo
<img width="644" alt="Screenshot 2023-07-25 at 11 11 31" src="https://github.com/pulgamecanica/42Portfolio/assets/79864209/54e0226b-5c2a-4616-8eb8-0a5baf5600f4">
<img width="642" alt="Screenshot 2023-07-25 at 11 10 33" src="https://github.com/pulgamecanica/42Portfolio/assets/79864209/10e2f979-d2fc-4071-a9eb-edc0fcc75acf">
| 1.0 | Wiki has a minor typo - I tried to create a PR, but I don't have access permissions to the wiki repo
<img width="644" alt="Screenshot 2023-07-25 at 11 11 31" src="https://github.com/pulgamecanica/42Portfolio/assets/79864209/54e0226b-5c2a-4616-8eb8-0a5baf5600f4">
<img width="642" alt="Screenshot 2023-07-25 at 11 10 33" src="https://github.com/pulgamecanica/42Portfolio/assets/79864209/10e2f979-d2fc-4071-a9eb-edc0fcc75acf">
| non_defect | wiki has a minor typo i tried to create a pr but i don t have access permissions to the wiki repo img width alt screenshot at src img width alt screenshot at src | 0 |
53,416 | 11,049,152,828 | IssuesEvent | 2019-12-09 22:52:04 | Chicago/design-cds-bootstrap | https://api.github.com/repos/Chicago/design-cds-bootstrap | closed | Make an example landing page | code |
- [x] header
- [x] hero or carousel
- [x] other content
- [x] footer | 1.0 | Make an example landing page -
- [x] header
- [x] hero or carousel
- [x] other content
- [x] footer | non_defect | make an example landing page header hero or carousel other content footer | 0 |
53,536 | 13,261,837,482 | IssuesEvent | 2020-08-20 20:37:21 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [RELEASE] IceRec L3 Muon (Trac #1622) | Migrated from Trac combo reconstruction defect | New release for IC2015-L3_Muon_V1 required to process runs with scintillators.
The STTools need to be updated like in L2.
The level3_Master.py needs an update to avoid a failed job when having an empty i3 file.
Two new files are required:
* http://code.icecube.wisc.edu/projects/icecube/changeset/143682/IceCube/projects/level3-filter-muon/trunk/python/level3_Master.py
* http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/STTools/trunk/python/seededRT/configuration_services.py
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1622">https://code.icecube.wisc.edu/projects/icecube/ticket/1622</a>, reported by joertlinand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-04-05T19:38:16",
"_ts": "1459885096502199",
"description": "New release for IC2015-L3_Muon_V1 required to process runs with scintillators.\n\nThe STTools need to be updated like in L2.\nThe level3_Master.py needs an update to avoid a failed job when having an empty i3 file.\n\nTwo new files are required:\n* http://code.icecube.wisc.edu/projects/icecube/changeset/143682/IceCube/projects/level3-filter-muon/trunk/python/level3_Master.py\n* http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/STTools/trunk/python/seededRT/configuration_services.py",
"reporter": "joertlin",
"cc": "",
"resolution": "fixed",
"time": "2016-04-05T18:59:54",
"component": "combo reconstruction",
"summary": "[RELEASE] IceRec L3 Muon",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [RELEASE] IceRec L3 Muon (Trac #1622) - New release for IC2015-L3_Muon_V1 required to process runs with scintillators.
The STTools need to be updated like in L2.
The level3_Master.py needs an update to avoid a failed job when having an empty i3 file.
Two new files are required:
* http://code.icecube.wisc.edu/projects/icecube/changeset/143682/IceCube/projects/level3-filter-muon/trunk/python/level3_Master.py
* http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/STTools/trunk/python/seededRT/configuration_services.py
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1622">https://code.icecube.wisc.edu/projects/icecube/ticket/1622</a>, reported by joertlinand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-04-05T19:38:16",
"_ts": "1459885096502199",
"description": "New release for IC2015-L3_Muon_V1 required to process runs with scintillators.\n\nThe STTools need to be updated like in L2.\nThe level3_Master.py needs an update to avoid a failed job when having an empty i3 file.\n\nTwo new files are required:\n* http://code.icecube.wisc.edu/projects/icecube/changeset/143682/IceCube/projects/level3-filter-muon/trunk/python/level3_Master.py\n* http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/STTools/trunk/python/seededRT/configuration_services.py",
"reporter": "joertlin",
"cc": "",
"resolution": "fixed",
"time": "2016-04-05T18:59:54",
"component": "combo reconstruction",
"summary": "[RELEASE] IceRec L3 Muon",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| defect | icerec muon trac new release for muon required to process runs with scintillators the sttools need to be updated like in the master py needs an update to avoid a failed job when having an empty file two new files are required migrated from json status closed changetime ts description new release for muon required to process runs with scintillators n nthe sttools need to be updated like in nthe master py needs an update to avoid a failed job when having an empty file n ntwo new files are required n reporter joertlin cc resolution fixed time component combo reconstruction summary icerec muon priority blocker keywords milestone owner olivas type defect | 1 |
33,333 | 7,093,333,593 | IssuesEvent | 2018-01-12 20:02:58 | CocoaPods/CocoaPods | https://api.github.com/repos/CocoaPods/CocoaPods | closed | Two different sources providing the same pod yields wrong resolved version | d3:hard s2:confirmed t2:defect | This is a link to https://github.com/CocoaPods/Molinillo/issues/88.
The summary is that the resolution of a pod that exists in at least two sources does not resolve properly.
It is unknown whether the bug is in the integration of CocoaPods and Molinillo 0.6.x or in Molinillo itself.
I am filing this issue in this project because an outside consumer of CocoaPods will not get the behavior expected when using CocoaPods master (or 1.4.x when it ships). | 1.0 | Two different sources providing the same pod yields wrong resolved version - This is a link to https://github.com/CocoaPods/Molinillo/issues/88.
The summary is that the resolution of a pod that exists in at least two sources does not resolve properly.
It is unknown whether the bug is in the integration of CocoaPods and Molinillo 0.6.x or in Molinillo itself.
I am filing this issue in this project because an outside consumer of CocoaPods will not get the behavior expected when using CocoaPods master (or 1.4.x when it ships). | defect | two different sources providing the same pod yields wrong resolved version this is a link to the summary is that the resolution of a pod that exists in at least two sources does not resolve properly it is unknown whether the bug is in the integration of cocoapods and molinillo x or in molinillo itself i am filing this issue in this project because an outside consumer of cocoapods will not get the behavior expected when using cocoapods master or x when it ships | 1 |
77,674 | 27,104,304,128 | IssuesEvent | 2023-02-15 10:56:07 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | DataTable: Navigation from header facet causes the value EL to be re-evaluated | :lady_beetle: defect | ### Describe the bug
Navigation from header facet causes the value EL to be re-evaluated, recreating the bean referenced by the EL in a new scope. This recreation is unnecessary as is for no use and may cause problems. This doesn't happen for navigation from table rows.
### Reproducer
[primefaces-test-master.zip](https://github.com/primefaces/primefaces/files/10725355/primefaces-test-master.zip)
1. Click the menu item to navigate to new page.
2. Both webbeans are reconstructed. The current one shouldn't be.
### Expected behavior
The current webbean should not be reconstructed after navigation, as the current scope is no longer valid.
### PrimeFaces edition
None
### PrimeFaces version
12
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3.19
### Java version
1.8
### Browser(s)
_No response_ | 1.0 | DataTable: Navigation from header facet causes the value EL to be re-evaluated - ### Describe the bug
Navigation from header facet causes the value EL to be re-evaluated, recreating the bean referenced by the EL in a new scope. This recreation is unnecessary as is for no use and may cause problems. This doesn't happen for navigation from table rows.
### Reproducer
[primefaces-test-master.zip](https://github.com/primefaces/primefaces/files/10725355/primefaces-test-master.zip)
1. Click the menu item to navigate to new page.
2. Both webbeans are reconstructed. The current one shouldn't be.
### Expected behavior
The current webbean should not be reconstructed after navigation, as the current scope is no longer valid.
### PrimeFaces edition
None
### PrimeFaces version
12
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3.19
### Java version
1.8
### Browser(s)
_No response_ | defect | datatable navigation from header facet causes the value el to be re evaluated describe the bug navigation from header facet causes the value el to be re evaluated recreating the bean referenced by the el in a new scope this recreation is unnecessary as is for no use and may cause problems this doesn t happen for navigation from table rows reproducer click the menu item to navigate to new page both webbeans are reconstructed the current one shouldn t be expected behavior the current webbean should not be reconstructed after navigation as the current scope is no longer valid primefaces edition none primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response | 1 |
4,980 | 2,610,163,182 | IssuesEvent | 2015-02-26 18:51:45 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | Graphics Glitch | auto-migrated Priority-Medium Type-Defect | ```
Hailfire droid right wheel has quite a wobble to it
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 3 May 2011 at 6:29 | 1.0 | Graphics Glitch - ```
Hailfire droid right wheel has quite a wobble to it
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 3 May 2011 at 6:29 | defect | graphics glitch hailfire droid right wheel has quite a wobble to it original issue reported on code google com by gmail com on may at | 1 |
57,793 | 16,063,668,977 | IssuesEvent | 2021-04-23 15:45:24 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | 508-defect-3 [COGNITION]: Edit, cancel, and remove buttons SHOULD be styled as buttons | 508-defect-3 508-issue-semantic-markup 508/Accessibility design direct deposit needs-grooming planned-work profile staging-review vsa vsa-authenticated-exp | # [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
## Feedback framework
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Definition of done
1. Review and acknowledge feedback.
1. Fix and/or document decisions made.
1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
<hr/>
## Point of Contact
**VFS Point of Contact:** Josh Kim
## Details
The "edit" buttons on this page are styled as links which is materially dishonest causing several issues:
- sighted users may think the "editing" action will occur on a new page
- voice interface users may not be able to activate the button should they say "click on edit link"
- screen reader users may be confused on help calls if they are instructed to click on an edit link when the screen reader announces it as a button
- low vision screen reader users may be confused that the screen reader is announcing a visually styled link as a button
- breaks consistency with other edit buttons across the interface
"Cancel" buttons on this page may face less of the issues listed above as they may be less likely to be misinterpreted as a link (e.g. hard to picture "cancelling" something on another page vs. editing)-- however, they still may be prone to misinterpretation and are still materially dishonest.
For more on this long-tail effort, check out our [buttons vs. links Mural board](https://app.mural.co/t/vsa8243/m/vsa8243/1610052905994/253723ef4b2ee9a726f538876a257df94d468db1) and this [Slack thread discussing _why_ material honesty is so important](https://dsva.slack.com/archives/C8E985R32/p1610659601009200?thread_ts=1610652897.002600&cid=C8E985R32).
"A few years at an accessibility event, a blind man was telling a story about how he was on the phone with someone who was trying to help him sign up for a service. The person on the phone told him to click the “sign up button” and he couldn’t find it because it was a link. He told the story with a lot more gravity than I can, but that was the point where I realized how important the human-human communication part of accessibility is, it’s not just machine to human." - @timwright12
## Acceptance Criteria
- [ ] "Edit" buttons are styled as buttons
- [ ] "Cancel" buttons are styled as buttons
## Environment
* Operating System: all
* Browser: any
* Screenreading device: any
* Server destination: staging
## Steps to Recreate
1. Enter `https://staging.va.gov/profile/direct-deposit` in browser
2. Inspect "edit"
3. Confirm it's a button
## Solution (if known)
Style the link as a button, options may include:
- Using primary or secondary outline button styles
- Using an icon + text pairing
## WCAG or Vendor Guidance (optional)
* [Adam Silver: But sometimes buttons look like links](https://adamsilver.io/articles/but-sometimes-buttons-look-like-links/)
* [VA Design System guidance on buttons](https://design.va.gov/components/buttons#guidance)
## Screenshots or Trace Logs
<img width="1667" alt="Screen Shot 2021-02-05 at 1 10 10 PM" src="https://user-images.githubusercontent.com/14154792/107072766-72a32200-67b4-11eb-85ab-7cc2afd60ac8.png">
| 1.0 | 508-defect-3 [COGNITION]: Edit, cancel, and remove buttons SHOULD be styled as buttons - # [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
## Feedback framework
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Definition of done
1. Review and acknowledge feedback.
1. Fix and/or document decisions made.
1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
<hr/>
## Point of Contact
**VFS Point of Contact:** Josh Kim
## Details
The "edit" buttons on this page are styled as links which is materially dishonest causing several issues:
- sighted users may think the "editing" action will occur on a new page
- voice interface users may not be able to activate the button should they say "click on edit link"
- screen reader users may be confused on help calls if they are instructed to click on an edit link when the screen reader announces it as a button
- low vision screen reader users may be confused that the screen reader is announcing a visually styled link as a button
- breaks consistency with other edit buttons across the interface
"Cancel" buttons on this page may face less of the issues listed above as they may be less likely to be misinterpreted as a link (e.g. hard to picture "cancelling" something on another page vs. editing)-- however, they still may be prone to misinterpretation and are still materially dishonest.
For more on this long-tail effort, check out our [buttons vs. links Mural board](https://app.mural.co/t/vsa8243/m/vsa8243/1610052905994/253723ef4b2ee9a726f538876a257df94d468db1) and this [Slack thread discussing _why_ material honesty is so important](https://dsva.slack.com/archives/C8E985R32/p1610659601009200?thread_ts=1610652897.002600&cid=C8E985R32).
"A few years at an accessibility event, a blind man was telling a story about how he was on the phone with someone who was trying to help him sign up for a service. The person on the phone told him to click the “sign up button” and he couldn’t find it because it was a link. He told the story with a lot more gravity than I can, but that was the point where I realized how important the human-human communication part of accessibility is, it’s not just machine to human." - @timwright12
## Acceptance Criteria
- [ ] "Edit" buttons are styled as buttons
- [ ] "Cancel" buttons are styled as buttons
## Environment
* Operating System: all
* Browser: any
* Screenreading device: any
* Server destination: staging
## Steps to Recreate
1. Enter `https://staging.va.gov/profile/direct-deposit` in browser
2. Inspect "edit"
3. Confirm it's a button
## Solution (if known)
Style the link as a button, options may include:
- Using primary or secondary outline button styles
- Using an icon + text pairing
## WCAG or Vendor Guidance (optional)
* [Adam Silver: But sometimes buttons look like links](https://adamsilver.io/articles/but-sometimes-buttons-look-like-links/)
* [VA Design System guidance on buttons](https://design.va.gov/components/buttons#guidance)
## Screenshots or Trace Logs
<img width="1667" alt="Screen Shot 2021-02-05 at 1 10 10 PM" src="https://user-images.githubusercontent.com/14154792/107072766-72a32200-67b4-11eb-85ab-7cc2afd60ac8.png">
| defect | defect edit cancel and remove buttons should be styled as buttons feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact josh kim details the edit buttons on this page are styled as links which is materially dishonest causing several issues sighted users may think the editing action will occur on a new page voice interface users may not be able to activate the button should they say click on edit link screen reader users may be confused on help calls if they are instructed to click on an edit link when the screen reader announces it as a button low vision screen reader users may be confused that the screen reader is announcing a visually styled link as a button breaks consistency with other edit buttons across the interface cancel buttons on this page may face less of the issues listed above as they may be less likely to be misinterpreted as a link e g hard to picture cancelling something on another page vs editing however they still may be prone to misinterpretation and are still materially dishonest for more on this long tail effort check out our and this a few years at an accessibility event a blind man was telling a story about how he was on the phone with someone who was trying to help him sign up for a service the person on the phone told him to click the “sign up button” and he couldn’t find it because it was a link he told the story with a lot more gravity than i can but that was the point where i realized how important the human human communication part of accessibility is it’s not just machine to human acceptance criteria edit buttons are styled as buttons cancel buttons are styled as buttons environment operating system all browser any screenreading device any server destination staging steps to recreate enter in browser inspect edit confirm it s a button solution if known style the link as a button options may include using primary or secondary outline button styles using an icon text pairing wcag or vendor guidance optional screenshots or trace logs img width alt screen shot at pm src | 1 |
69,163 | 22,238,692,338 | IssuesEvent | 2022-06-09 01:09:37 | Cockatrice/Cockatrice | https://api.github.com/repos/Cockatrice/Cockatrice | closed | 64 bit builds no longer work on windows 7 because of qt6 | OS - Windows Defect - Regression Bug High Priority | Client Version: 2.8.1-beta.6 (2022-01-16)
Client Operating System: Windows 7 Version 6.1 (Build 7601: SP 1)
Build Architecture: 64-bit
Qt Version: 5.15.2
System Locale: en_US
Install Mode: Standard
_______________________________________________________________________________________
Install beta.7. When I run cockatrice I get Windows Application Error "The application was unable to start correctly (0xc0000005)"
Try install beta.8. Same thing.
Try install beta.9. Same thing.
Restore beta.6. Everything is fine again.
| 1.0 | 64 bit builds no longer work on windows 7 because of qt6 - Client Version: 2.8.1-beta.6 (2022-01-16)
Client Operating System: Windows 7 Version 6.1 (Build 7601: SP 1)
Build Architecture: 64-bit
Qt Version: 5.15.2
System Locale: en_US
Install Mode: Standard
_______________________________________________________________________________________
Install beta.7. When I run cockatrice I get Windows Application Error "The application was unable to start correctly (0xc0000005)"
Try install beta.8. Same thing.
Try install beta.9. Same thing.
Restore beta.6. Everything is fine again.
| defect | bit builds no longer work on windows because of client version beta client operating system windows version build sp build architecture bit qt version system locale en us install mode standard install beta when i run cockatrice i get windows application error the application was unable to start correctly try install beta same thing try install beta same thing restore beta everything is fine again | 1 |
39,563 | 9,547,788,089 | IssuesEvent | 2019-05-02 01:15:28 | idaholab/moose | https://api.github.com/repos/idaholab/moose | opened | Some MeshGenerator tests need exodiff to be pedantic | T: defect | ## Bug Description
Mesh generators testing for making sidesets and nodesets will pass tests in a false positive manner. These tests should run in the harness with `exodiff_opts = '-pedantic'` so exodiff reports different meshes and the test fails.
## Steps to Reproduce
Take `moose/test/tests/meshgenerators/sidesets_between_subdomains_generator/sideset_between_subdomains.i` and comment out the `central_boundary` block which is what is being tested, run the test and it passes!
I'm not sure how many of the MeshGenerator tests would fail this way, but likely worth it to add the exodiff_opts to all of those tests.
## Impact
Make some MeshGenerator tests actually test the things they say they are testing.
| 1.0 | Some MeshGenerator tests need exodiff to be pedantic - ## Bug Description
Mesh generators testing for making sidesets and nodesets will pass tests in a false positive manner. These tests should run in the harness with `exodiff_opts = '-pedantic'` so exodiff reports different meshes and the test fails.
## Steps to Reproduce
Take `moose/test/tests/meshgenerators/sidesets_between_subdomains_generator/sideset_between_subdomains.i` and comment out the `central_boundary` block which is what is being tested, run the test and it passes!
I'm not sure how many of the MeshGenerator tests would fail this way, but likely worth it to add the exodiff_opts to all of those tests.
## Impact
Make some MeshGenerator tests actually test the things they say they are testing.
| defect | some meshgenerator tests need exodiff to be pedantic bug description mesh generators testing for making sidesets and nodesets will pass tests in a false positive manner these tests should run in the harness with exodiff opts pedantic so exodiff reports different meshes and the test fails steps to reproduce take moose test tests meshgenerators sidesets between subdomains generator sideset between subdomains i and comment out the central boundary block which is what is being tested run the test and it passes i m not sure how many of the meshgenerator tests would fail this way but likely worth it to add the exodiff opts to all of those tests impact make some meshgenerator tests actually test the things they say they are testing | 1 |
29,726 | 5,843,807,598 | IssuesEvent | 2017-05-10 10:06:00 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | WAN backup events are published twice | Team: Core Type: Defect | The `publishWanReplicationEventBackups` method is called in the `run` and `afterRun` method in the `PartitionWideEntryBackupOperation`. Elsewhere it is only called in `afterRun`. This needs to be checked but possibly the fix is just to remove the call in `run`. | 1.0 | WAN backup events are published twice - The `publishWanReplicationEventBackups` method is called in the `run` and `afterRun` method in the `PartitionWideEntryBackupOperation`. Elsewhere it is only called in `afterRun`. This needs to be checked but possibly the fix is just to remove the call in `run`. | defect | wan backup events are published twice the publishwanreplicationeventbackups method is called in the run and afterrun method in the partitionwideentrybackupoperation elsewhere it is only called in afterrun this needs to be checked but possibly the fix is just to remove the call in run | 1 |
86,215 | 8,029,168,454 | IssuesEvent | 2018-07-27 15:10:03 | smap-consulting/smapserver | https://api.github.com/repos/smap-consulting/smapserver | closed | ENHANCEMENT - Include 'not in' as filtering option in search | In Test enhancement | ### Component fieldTask
### Expected Behaviour: 'in' filtering option is a great feature to use in search like search('file','in','column',${subset}). Maybe you could consider to add 'not in' to have a chance to exclude
like search('file','not in','column',${subset}) | 1.0 | ENHANCEMENT - Include 'not in' as filtering option in search - ### Component fieldTask
### Expected Behaviour: 'in' filtering option is a great feature to use in search like search('file','in','column',${subset}). Maybe you could consider to add 'not in' to have a chance to exclude
like search('file','not in','column',${subset}) | non_defect | enhancement include not in as filtering option in search component fieldtask expected behaviour in filtering option is a great feature to use in search like search file in column subset maybe you could consider to add not in to have a chance to exclude like search file not in column subset | 0 |
2,562 | 2,607,928,254 | IssuesEvent | 2015-02-26 00:25:36 | chrsmithdemos/minify | https://api.github.com/repos/chrsmithdemos/minify | closed | Builder app leaves "?" in group URI | auto-migrated Priority-Medium Release-2.1.0 Type-Defect | ```
When the builder verifies that mod_rewrite is working, it removes the "?"
from @import "/min/?g=css2"; but not from /min/?g=keyName
```
-----
Original issue reported on code.google.com by `mrclay....@gmail.com` on 20 Sep 2008 at 2:54 | 1.0 | Builder app leaves "?" in group URI - ```
When the builder verifies that mod_rewrite is working, it removes the "?"
from @import "/min/?g=css2"; but not from /min/?g=keyName
```
-----
Original issue reported on code.google.com by `mrclay....@gmail.com` on 20 Sep 2008 at 2:54 | defect | builder app leaves in group uri when the builder verifies that mod rewrite is working it removes the from import min g but not from min g keyname original issue reported on code google com by mrclay gmail com on sep at | 1 |
56,904 | 15,438,134,195 | IssuesEvent | 2021-03-07 19:16:00 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | Wrong results from `capped_distance()` with "nsgrid" search and atom in the center of the box | Component-lib PBC Priority-High defect | ## Expected behavior ##
<!-- A clear and concise description of what you want to do and what you think should happen. (Code to reproduce the behavior can be added below). -->
`capped_distance()` function should always give the correct result regardless of the employed search method and atomic positions.
In the example below the expected result for cutoff=3.2 is:
* cutoff=3.2
bruteforce: 2497 pairs
pkdtree: 2497 pairs
nsgrid: 2497 pairs
## Actual behavior ##
<!-- What happened instead. Add as much detail as you can. Include (copy and paste) stack traces and any output. -->
When one of the atoms of the reference group is positioned exactly in the center of the box, `capped_distance()` sometimes returns erroneous results for "nsgrid" search method.
* cutoff=2.8
bruteforce: 1115 pairs
pkdtree: 1115 pairs
nsgrid: 1115 pairs
* cutoff=3.2
bruteforce: 2497 pairs
pkdtree: 2497 pairs
nsgrid: 2510 pairs
## Code to reproduce the behavior ##
<!-- Show us how to reproduce the failure. If you can, use trajectory files from the test data. Use the code snipped below as a starting point. -->
``` python
import MDAnalysis as mda
from MDAnalysis.lib.distances import capped_distance
from MDAnalysis.transformations.translate import center_in_box
from MDAnalysis.tests.datafiles import PDB_xvf
u = mda.Universe(PDB_xvf)
ag = u.select_atoms('index 0')
u.trajectory.ts = center_in_box(ag)(u.trajectory.ts)
box = u.dimensions
reference = u.select_atoms('protein')
configuration = u.select_atoms('not protein')
for cutoff in [2.8, 3.2]:
print(f"* cutoff={cutoff}")
for method in ['bruteforce', 'pkdtree', 'nsgrid']:
pairs, distances = capped_distance(
reference.positions,
configuration.positions,
max_cutoff=cutoff,
box=box,
method=method,
return_distances=True,
)
print(f"{method}: {len(pairs)} pairs")
```
## Current version of MDAnalysis ##
- MDAnalysis 1.0.0 (conda-forge)
- Python 3.7.7
- Ubuntu 18.04.5 LTS
| 1.0 | Wrong results from `capped_distance()` with "nsgrid" search and atom in the center of the box - ## Expected behavior ##
<!-- A clear and concise description of what you want to do and what you think should happen. (Code to reproduce the behavior can be added below). -->
`capped_distance()` function should always give the correct result regardless of the employed search method and atomic positions.
In the example below the expected result for cutoff=3.2 is:
* cutoff=3.2
bruteforce: 2497 pairs
pkdtree: 2497 pairs
nsgrid: 2497 pairs
## Actual behavior ##
<!-- What happened instead. Add as much detail as you can. Include (copy and paste) stack traces and any output. -->
When one of the atoms of the reference group is positioned exactly in the center of the box, `capped_distance()` sometimes returns erroneous results for "nsgrid" search method.
* cutoff=2.8
bruteforce: 1115 pairs
pkdtree: 1115 pairs
nsgrid: 1115 pairs
* cutoff=3.2
bruteforce: 2497 pairs
pkdtree: 2497 pairs
nsgrid: 2510 pairs
## Code to reproduce the behavior ##
<!-- Show us how to reproduce the failure. If you can, use trajectory files from the test data. Use the code snipped below as a starting point. -->
``` python
import MDAnalysis as mda
from MDAnalysis.lib.distances import capped_distance
from MDAnalysis.transformations.translate import center_in_box
from MDAnalysis.tests.datafiles import PDB_xvf
u = mda.Universe(PDB_xvf)
ag = u.select_atoms('index 0')
u.trajectory.ts = center_in_box(ag)(u.trajectory.ts)
box = u.dimensions
reference = u.select_atoms('protein')
configuration = u.select_atoms('not protein')
for cutoff in [2.8, 3.2]:
print(f"* cutoff={cutoff}")
for method in ['bruteforce', 'pkdtree', 'nsgrid']:
pairs, distances = capped_distance(
reference.positions,
configuration.positions,
max_cutoff=cutoff,
box=box,
method=method,
return_distances=True,
)
print(f"{method}: {len(pairs)} pairs")
```
## Current version of MDAnalysis ##
- MDAnalysis 1.0.0 (conda-forge)
- Python 3.7.7
- Ubuntu 18.04.5 LTS
| defect | wrong results from capped distance with nsgrid search and atom in the center of the box expected behavior capped distance function should always give the correct result regardless of the employed search method and atomic positions in the example below the expected result for cutoff is cutoff bruteforce pairs pkdtree pairs nsgrid pairs actual behavior when one of the atoms of the reference group is positioned exactly in the center of the box capped distance sometimes returns erroneous results for nsgrid search method cutoff bruteforce pairs pkdtree pairs nsgrid pairs cutoff bruteforce pairs pkdtree pairs nsgrid pairs code to reproduce the behavior python import mdanalysis as mda from mdanalysis lib distances import capped distance from mdanalysis transformations translate import center in box from mdanalysis tests datafiles import pdb xvf u mda universe pdb xvf ag u select atoms index u trajectory ts center in box ag u trajectory ts box u dimensions reference u select atoms protein configuration u select atoms not protein for cutoff in print f cutoff cutoff for method in pairs distances capped distance reference positions configuration positions max cutoff cutoff box box method method return distances true print f method len pairs pairs current version of mdanalysis mdanalysis conda forge python ubuntu lts | 1 |
184,204 | 31,838,188,907 | IssuesEvent | 2023-09-14 14:41:15 | alor-broker/Astras-Trading-UI | https://api.github.com/repos/alor-broker/Astras-Trading-UI | opened | [FEATURE] ... Переименовать поддержку в Телеграм | enhancement recently reported design | ### Please describe, is your feature request related to a problem?
Сейчас непонятно, что пункт "Поддержка" подразумевает переход в Телеграм

### Describe the solution you would like
Переименовать на "Телеграм"
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct and Contribution guideline | 1.0 | [FEATURE] ... Переименовать поддержку в Телеграм - ### Please describe, is your feature request related to a problem?
Сейчас непонятно, что пункт "Поддержка" подразумевает переход в Телеграм

### Describe the solution you would like
Переименовать на "Телеграм"
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct and Contribution guideline | non_defect | переименовать поддержку в телеграм please describe is your feature request related to a problem сейчас непонятно что пункт поддержка подразумевает переход в телеграм describe the solution you would like переименовать на телеграм describe alternatives you ve considered no response additional context no response code of conduct i agree to follow this project s code of conduct and contribution guideline | 0 |
313,852 | 23,494,386,612 | IssuesEvent | 2022-08-17 22:33:58 | mszostok/version | https://api.github.com/repos/mszostok/version | closed | Add Go playground badges | area/documentation enhancement | **Description**
Create a dedicated Go playground snippets for all examples and update documentation badges.
Example view:
## [Custom Layout](https://github.com/mszostok/version/tree/main/example/custom-layout/main.go) [![playground][play-simple-img]][play-simple]

```bash
# Build
go build -ldflags "-X 'github.com/mszostok/version.buildDate=`date`'" -o example ./custom-layout
# Showcase
./example
```
[play-simple-img]: https://img.shields.io/badge/playground-custom-29BEB0?style=flat&logo=go
[play-simple]: https://go.dev/play/p/H1GufyrAw3T
| 1.0 | Add Go playground badges - **Description**
Create a dedicated Go playground snippets for all examples and update documentation badges.
Example view:
## [Custom Layout](https://github.com/mszostok/version/tree/main/example/custom-layout/main.go) [![playground][play-simple-img]][play-simple]

```bash
# Build
go build -ldflags "-X 'github.com/mszostok/version.buildDate=`date`'" -o example ./custom-layout
# Showcase
./example
```
[play-simple-img]: https://img.shields.io/badge/playground-custom-29BEB0?style=flat&logo=go
[play-simple]: https://go.dev/play/p/H1GufyrAw3T
| non_defect | add go playground badges description create a dedicated go playground snippets for all examples and update documentation badges example view bash build go build ldflags x github com mszostok version builddate date o example custom layout showcase example | 0 |
33,338 | 7,093,852,655 | IssuesEvent | 2018-01-12 22:24:14 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | MooseDocs can get tangled up in libMesh submodule content | C: Documentation P: minor T: defect | ## Rationale
<!--What is the reason for this enhancement or what error are you reporting?-->
If you are using the libMesh submodule it is possible that MooseDocs, when searching
for class definitions in the source code, can have problems. In addition, searching
libMesh source code is not needed.
## Description
<!--Provide details of the enhancement or instructions for reproducing the error.-->
I can't reproduce it, but see the problem when walking over all directories in MooseClassDatabase.
## Impact
<!--How will the changes impact the code, developers, and users? Discuss changes to the
internal interfaces and public API.-->
Internal change only, the MooseClassDatabase just needs to be restricted. | 1.0 | MooseDocs can get tangled up in libMesh submodule content - ## Rationale
<!--What is the reason for this enhancement or what error are you reporting?-->
If you are using the libMesh submodule it is possible that MooseDocs, when searching
for class definitions in the source code, can have problems. In addition, searching
libMesh source code is not needed.
## Description
<!--Provide details of the enhancement or instructions for reproducing the error.-->
I can't reproduce it, but see the problem when walking over all directories in MooseClassDatabase.
## Impact
<!--How will the changes impact the code, developers, and users? Discuss changes to the
internal interfaces and public API.-->
Internal change only, the MooseClassDatabase just needs to be restricted. | defect | moosedocs can get tangled up in libmesh submodule content rationale if you are using the libmesh submodule it is possible that moosedocs when searching for class definitions in the source code can have problems in addition searching libmesh source code is not needed description i can t reproduce it but see the problem when walking over all directories in mooseclassdatabase impact how will the changes impact the code developers and users discuss changes to the internal interfaces and public api internal change only the mooseclassdatabase just needs to be restricted | 1 |
417,428 | 28,110,479,589 | IssuesEvent | 2023-03-31 06:42:59 | kynhan/ped | https://api.github.com/repos/kynhan/ped | opened | Find command not in command summary in UG | severity.Low type.DocumentationBug | User guide does not have a row for the Find command in the command summary at the bottom.

<!--session: 1680242554955-bf889941-44cc-4197-8386-7318ab22a3ac-->
<!--Version: Web v3.4.7--> | 1.0 | Find command not in command summary in UG - User guide does not have a row for the Find command in the command summary at the bottom.

<!--session: 1680242554955-bf889941-44cc-4197-8386-7318ab22a3ac-->
<!--Version: Web v3.4.7--> | non_defect | find command not in command summary in ug user guide does not have a row for the find command in the command summary at the bottom | 0 |
181,491 | 14,040,561,666 | IssuesEvent | 2020-11-01 03:23:57 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | airgap - no fleet agents in downstream clusters after upgrading from 2.4.8 to 2.5.2-rc5 | [zube]: To Test kind/bug-qa status/blocker | <!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
- run rancher:v2.4.8 single-install in the airgap env
- add a couple of clusters
- upgrade rancher to v2.5.2-rc5
**Result:**
- Clusters have no nodes showing in the dashboard

- the bundle `fleet-agent` is waiting for applying

- and there is no deployment for fleet-agent in the downstream clusters
**Other details that may be helpful:**
- in the local cluster the fleet agent and controller are deployed and running

- the logs of the `fleet-controller` pod in the local cluster
<details>
<summary>Details</summary>
time="2020-11-01T01:05:54Z" level=info msg="No access to list CRDs, assuming CRDs are pre-created."
I1101 01:05:54.713261 1 leaderelection.go:242] attempting to acquire leader lease fleet-system/fleet-controller-lock...
I1101 01:05:54.718411 1 leaderelection.go:252] successfully acquired lease fleet-system/fleet-controller-lock
E1101 01:05:54.747222 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleNamespaceMapping: the server could not find the requested resource (get bundlenamespacemappings.meta.k8s.io)
E1101 01:05:54.748243 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterGroup: the server could not find the requested resource (get clustergroups.meta.k8s.io)
E1101 01:05:54.748550 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistration: the server could not find the requested resource (get clusterregistrations.meta.k8s.io)
E1101 01:05:54.751996 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistrationToken: the server could not find the requested resource (get clusterregistrationtokens.meta.k8s.io)
E1101 01:05:54.791021 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Cluster: the server could not find the requested resource (get clusters.meta.k8s.io)
E1101 01:05:54.803015 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Bundle: the server could not find the requested resource (get bundles.meta.k8s.io)
E1101 01:05:54.803059 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Content: the server could not find the requested resource (get contents.meta.k8s.io)
E1101 01:05:54.812784 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepo: the server could not find the requested resource (get gitrepos.meta.k8s.io)
E1101 01:05:54.827065 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleDeployment: the server could not find the requested resource (get bundledeployments.meta.k8s.io)
E1101 01:05:54.827428 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1.GitJob: the server could not find the requested resource (get gitjobs.meta.k8s.io)
E1101 01:05:54.828093 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepoRestriction: the server could not find the requested resource (get gitreporestrictions.meta.k8s.io)
E1101 01:05:55.722299 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterGroup: the server could not find the requested resource (get clustergroups.meta.k8s.io)
E1101 01:05:55.776879 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepo: the server could not find the requested resource (get gitrepos.meta.k8s.io)
E1101 01:05:55.807980 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistrationToken: the server could not find the requested resource (get clusterregistrationtokens.meta.k8s.io)
E1101 01:05:55.839297 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Bundle: the server could not find the requested resource (get bundles.meta.k8s.io)
E1101 01:05:55.856042 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistration: the server could not find the requested resource (get clusterregistrations.meta.k8s.io)
E1101 01:05:55.864206 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1.GitJob: the server could not find the requested resource (get gitjobs.meta.k8s.io)
E1101 01:05:55.877688 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepoRestriction: the server could not find the requested resource (get gitreporestrictions.meta.k8s.io)
E1101 01:05:55.967382 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Cluster: the server could not find the requested resource (get clusters.meta.k8s.io)
E1101 01:05:56.149797 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Content: the server could not find the requested resource (get contents.meta.k8s.io)
E1101 01:05:56.199477 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleNamespaceMapping: the server could not find the requested resource (get bundlenamespacemappings.meta.k8s.io)
E1101 01:05:56.318722 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleDeployment: the server could not find the requested resource (get bundledeployments.meta.k8s.io)
E1101 01:05:57.423269 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepo: the server could not find the requested resource (get gitrepos.meta.k8s.io)
E1101 01:05:57.593821 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1.GitJob: the server could not find the requested resource (get gitjobs.meta.k8s.io)
E1101 01:05:57.662592 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistrationToken: the server could not find the requested resource (get clusterregistrationtokens.meta.k8s.io)
E1101 01:05:57.663013 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Cluster: the server could not find the requested resource (get clusters.meta.k8s.io)
E1101 01:05:58.161796 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterGroup: the server could not find the requested resource (get clustergroups.meta.k8s.io)
E1101 01:05:58.197328 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleDeployment: the server could not find the requested resource (get bundledeployments.meta.k8s.io)
E1101 01:05:58.283489 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleNamespaceMapping: the server could not find the requested resource (get bundlenamespacemappings.meta.k8s.io)
E1101 01:05:58.413227 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Bundle: the server could not find the requested resource (get bundles.meta.k8s.io)
E1101 01:05:58.433382 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepoRestriction: the server could not find the requested resource (get gitreporestrictions.meta.k8s.io)
E1101 01:05:58.879993 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Content: the server could not find the requested resource (get contents.meta.k8s.io)
time="2020-11-01T01:06:05Z" level=info msg="All controllers have been started"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=BundleNamespaceMapping controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=ServiceAccount controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=GitRepo controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=Content controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=GitRepoRestriction controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=Namespace controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=BundleDeployment controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=ClusterRegistrationToken controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=ConfigMap controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRole controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=RoleBinding controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=ClusterRegistration controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=Bundle controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=ClusterGroup controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=Cluster controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting gitjob.cattle.io/v1, Kind=GitJob controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=Role controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=Secret controller"
time="2020-11-01T01:06:08Z" level=info msg="Deployed new agent for cluster fleet-local/local"
time="2020-11-01T01:06:10Z" level=info msg="Deleted old agent for cluster fleet-local/local"
time="2020-11-01T01:06:12Z" level=info msg="Deployed new agent for cluster fleet-local/local"
time="2020-11-01T01:06:17Z" level=info msg="Cluster registration fleet-local/request-hgtpt, cluster fleet-local/local granted [false]"
time="2020-11-01T01:06:17Z" level=info msg="Cluster registration fleet-local/request-hgtpt, cluster fleet-local/local granted [false]"
time="2020-11-01T01:06:17Z" level=info msg="Cluster registration fleet-local/request-hgtpt, cluster fleet-local/local granted [true]"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:40Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:11:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:11:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:12:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:13:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:13:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:14:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:16:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:21:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:21:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:22:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:22:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:22:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:23:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:26:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:26:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:27:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:27:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:27:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:30:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:32:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:32:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:32:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:37:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:37:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:41:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:41:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:41:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:42:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:42:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:42:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:42:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:46:17Z" level=info msg="Deleting expired registration secret fleet-clusters-system/c-0993e7869293b074613ec51b4ee2efe9f3ef0a86823e9b727683d9c90008d"
time="2020-11-01T01:46:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:46:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:46:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:50:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:50:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:51:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:51:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
</details>
| 1.0 | airgap - no fleet agents in downstream clusters after upgrading from 2.4.8 to 2.5.2-rc5 - <!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
- run rancher:v2.4.8 single-install in the airgap env
- add a couple of clusters
- upgrade rancher to v2.5.2-rc5
**Result:**
- Clusters have no nodes showing in the dashboard

- the bundle `fleet-agent` is waiting for applying

- and there is no deployment for fleet-agent in the downstream clusters
**Other details that may be helpful:**
- in the local cluster the fleet agent and controller are deployed and running

- the logs of the `fleet-controller` pod in the local cluster
<details>
<summary>Details</summary>
time="2020-11-01T01:05:54Z" level=info msg="No access to list CRDs, assuming CRDs are pre-created."
I1101 01:05:54.713261 1 leaderelection.go:242] attempting to acquire leader lease fleet-system/fleet-controller-lock...
I1101 01:05:54.718411 1 leaderelection.go:252] successfully acquired lease fleet-system/fleet-controller-lock
E1101 01:05:54.747222 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleNamespaceMapping: the server could not find the requested resource (get bundlenamespacemappings.meta.k8s.io)
E1101 01:05:54.748243 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterGroup: the server could not find the requested resource (get clustergroups.meta.k8s.io)
E1101 01:05:54.748550 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistration: the server could not find the requested resource (get clusterregistrations.meta.k8s.io)
E1101 01:05:54.751996 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistrationToken: the server could not find the requested resource (get clusterregistrationtokens.meta.k8s.io)
E1101 01:05:54.791021 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Cluster: the server could not find the requested resource (get clusters.meta.k8s.io)
E1101 01:05:54.803015 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Bundle: the server could not find the requested resource (get bundles.meta.k8s.io)
E1101 01:05:54.803059 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Content: the server could not find the requested resource (get contents.meta.k8s.io)
E1101 01:05:54.812784 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepo: the server could not find the requested resource (get gitrepos.meta.k8s.io)
E1101 01:05:54.827065 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleDeployment: the server could not find the requested resource (get bundledeployments.meta.k8s.io)
E1101 01:05:54.827428 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1.GitJob: the server could not find the requested resource (get gitjobs.meta.k8s.io)
E1101 01:05:54.828093 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepoRestriction: the server could not find the requested resource (get gitreporestrictions.meta.k8s.io)
E1101 01:05:55.722299 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterGroup: the server could not find the requested resource (get clustergroups.meta.k8s.io)
E1101 01:05:55.776879 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepo: the server could not find the requested resource (get gitrepos.meta.k8s.io)
E1101 01:05:55.807980 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistrationToken: the server could not find the requested resource (get clusterregistrationtokens.meta.k8s.io)
E1101 01:05:55.839297 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Bundle: the server could not find the requested resource (get bundles.meta.k8s.io)
E1101 01:05:55.856042 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistration: the server could not find the requested resource (get clusterregistrations.meta.k8s.io)
E1101 01:05:55.864206 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1.GitJob: the server could not find the requested resource (get gitjobs.meta.k8s.io)
E1101 01:05:55.877688 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepoRestriction: the server could not find the requested resource (get gitreporestrictions.meta.k8s.io)
E1101 01:05:55.967382 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Cluster: the server could not find the requested resource (get clusters.meta.k8s.io)
E1101 01:05:56.149797 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Content: the server could not find the requested resource (get contents.meta.k8s.io)
E1101 01:05:56.199477 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleNamespaceMapping: the server could not find the requested resource (get bundlenamespacemappings.meta.k8s.io)
E1101 01:05:56.318722 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleDeployment: the server could not find the requested resource (get bundledeployments.meta.k8s.io)
E1101 01:05:57.423269 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepo: the server could not find the requested resource (get gitrepos.meta.k8s.io)
E1101 01:05:57.593821 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1.GitJob: the server could not find the requested resource (get gitjobs.meta.k8s.io)
E1101 01:05:57.662592 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterRegistrationToken: the server could not find the requested resource (get clusterregistrationtokens.meta.k8s.io)
E1101 01:05:57.663013 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Cluster: the server could not find the requested resource (get clusters.meta.k8s.io)
E1101 01:05:58.161796 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.ClusterGroup: the server could not find the requested resource (get clustergroups.meta.k8s.io)
E1101 01:05:58.197328 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleDeployment: the server could not find the requested resource (get bundledeployments.meta.k8s.io)
E1101 01:05:58.283489 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.BundleNamespaceMapping: the server could not find the requested resource (get bundlenamespacemappings.meta.k8s.io)
E1101 01:05:58.413227 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Bundle: the server could not find the requested resource (get bundles.meta.k8s.io)
E1101 01:05:58.433382 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.GitRepoRestriction: the server could not find the requested resource (get gitreporestrictions.meta.k8s.io)
E1101 01:05:58.879993 1 reflector.go:178] pkg/mod/github.com/rancher/client-go@v0.18.8-fleet1/tools/cache/reflector.go:125: Failed to list *v1alpha1.Content: the server could not find the requested resource (get contents.meta.k8s.io)
time="2020-11-01T01:06:05Z" level=info msg="All controllers have been started"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=BundleNamespaceMapping controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=ServiceAccount controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=GitRepo controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=Content controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=GitRepoRestriction controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=Namespace controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=BundleDeployment controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=ClusterRegistrationToken controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=ConfigMap controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRole controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=RoleBinding controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=ClusterRegistration controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=Bundle controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=ClusterGroup controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting fleet.cattle.io/v1alpha1, Kind=Cluster controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting gitjob.cattle.io/v1, Kind=GitJob controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=Role controller"
time="2020-11-01T01:06:05Z" level=info msg="Starting /v1, Kind=Secret controller"
time="2020-11-01T01:06:08Z" level=info msg="Deployed new agent for cluster fleet-local/local"
time="2020-11-01T01:06:10Z" level=info msg="Deleted old agent for cluster fleet-local/local"
time="2020-11-01T01:06:12Z" level=info msg="Deployed new agent for cluster fleet-local/local"
time="2020-11-01T01:06:17Z" level=info msg="Cluster registration fleet-local/request-hgtpt, cluster fleet-local/local granted [false]"
time="2020-11-01T01:06:17Z" level=info msg="Cluster registration fleet-local/request-hgtpt, cluster fleet-local/local granted [false]"
time="2020-11-01T01:06:17Z" level=info msg="Cluster registration fleet-local/request-hgtpt, cluster fleet-local/local granted [true]"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:40Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:06:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:10Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:25Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:40Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:07:55Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:10Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:25Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:40Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:08:55Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:10Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:09:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:09:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:10:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:11:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:11:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:12:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:12:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:13:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:13:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:13:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:14:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:14:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:15:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:16:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:16:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:17:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:18:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:19:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:20:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:21:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:21:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:22:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:22:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:22:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:23:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:23:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:24:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:25:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:26:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:26:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:27:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:27:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:27:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:28:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:29:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:30:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:30:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:31:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:32:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:32:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:32:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:33:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:34:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:35:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:36:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:37:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:37:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:38:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:39:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:40:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:41:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:41:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:41:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": dial tcp 172.31.4.24:443: i/o timeout, requeuing"
time="2020-11-01T01:42:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:42:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:42:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:42:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:43:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:44:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:45:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:46:17Z" level=info msg="Deleting expired registration secret fleet-clusters-system/c-0993e7869293b074613ec51b4ee2efe9f3ef0a86823e9b727683d9c90008d"
time="2020-11-01T01:46:45Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:46:55Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:46:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:00Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:47:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:00Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:10Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:15Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:15Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:48:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:15Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:25Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:30Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:49:30Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:25Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:30Z" level=error msg="error syncing 'fleet-default/c-h45qq': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-h45qq/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:40Z" level=error msg="error syncing 'fleet-default/c-dlkxj': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-dlkxj/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:40Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:50:45Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded, requeuing"
time="2020-11-01T01:50:45Z" level=error msg="error syncing 'fleet-default/c-2rwdw': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2rwdw/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:50:55Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:51:00Z" level=error msg="error syncing 'fleet-default/c-l59ft': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-l59ft/version?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), requeuing"
time="2020-11-01T01:51:10Z" level=error msg="error syncing 'fleet-default/c-2dtrp': handler import-cluster: Get \"https://jiaqi320-37263-internal.qa.rancher.space/k8s/clusters/c-2dtrp/version?timeout=15s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers), requeuing"
</details>
| non_defect | airgap no fleet agents in downstream clusters after upgrading from to please search for existing issues first then read to see what we expect in an issue for security issues please email security rancher com instead of posting a public issue in github you may but are not required to use the gpg key located on keybase what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible run rancher single install in the airgap env add a couple of clusters upgrade rancher to result clusters have no nodes showing in the dashboard the bundle fleet agent is waiting for applying and there is no deployment for fleet agent in the downstream clusters other details that may be helpful in the local cluster the fleet agent and controller are deployed and running the logs of the fleet controller pod in the local cluster details time level info msg no access to list crds assuming crds are pre created leaderelection go attempting to acquire leader lease fleet system fleet controller lock leaderelection go successfully acquired lease fleet system fleet controller lock reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundlenamespacemapping the server could not find the requested resource get bundlenamespacemappings meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clustergroup the server could not find the requested resource get clustergroups meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clusterregistration the server could not find the requested resource get clusterregistrations meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clusterregistrationtoken the server could not find the requested resource get clusterregistrationtokens meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list cluster the server could not find the requested resource get clusters meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundle the server could not find the requested resource get bundles meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list content the server could not find the requested resource get contents meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitrepo the server could not find the requested resource get gitrepos meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundledeployment the server could not find the requested resource get bundledeployments meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitjob the server could not find the requested resource get gitjobs meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitreporestriction the server could not find the requested resource get gitreporestrictions meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clustergroup the server could not find the requested resource get clustergroups meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitrepo the server could not find the requested resource get gitrepos meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clusterregistrationtoken the server could not find the requested resource get clusterregistrationtokens meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundle the server could not find the requested resource get bundles meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clusterregistration the server could not find the requested resource get clusterregistrations meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitjob the server could not find the requested resource get gitjobs meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitreporestriction the server could not find the requested resource get gitreporestrictions meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list cluster the server could not find the requested resource get clusters meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list content the server could not find the requested resource get contents meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundlenamespacemapping the server could not find the requested resource get bundlenamespacemappings meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundledeployment the server could not find the requested resource get bundledeployments meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitrepo the server could not find the requested resource get gitrepos meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitjob the server could not find the requested resource get gitjobs meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clusterregistrationtoken the server could not find the requested resource get clusterregistrationtokens meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list cluster the server could not find the requested resource get clusters meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list clustergroup the server could not find the requested resource get clustergroups meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundledeployment the server could not find the requested resource get bundledeployments meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundlenamespacemapping the server could not find the requested resource get bundlenamespacemappings meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list bundle the server could not find the requested resource get bundles meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list gitreporestriction the server could not find the requested resource get gitreporestrictions meta io reflector go pkg mod github com rancher client go tools cache reflector go failed to list content the server could not find the requested resource get contents meta io time level info msg all controllers have been started time level info msg starting fleet cattle io kind bundlenamespacemapping controller time level info msg starting kind serviceaccount controller time level info msg starting fleet cattle io kind gitrepo controller time level info msg starting fleet cattle io kind content controller time level info msg starting fleet cattle io kind gitreporestriction controller time level info msg starting kind namespace controller time level info msg starting fleet cattle io kind bundledeployment controller time level info msg starting fleet cattle io kind clusterregistrationtoken controller time level info msg starting kind configmap controller time level info msg starting rbac authorization io kind clusterrole controller time level info msg starting rbac authorization io kind rolebinding controller time level info msg starting fleet cattle io kind clusterregistration controller time level info msg starting rbac authorization io kind clusterrolebinding controller time level info msg starting fleet cattle io kind bundle controller time level info msg starting fleet cattle io kind clustergroup controller time level info msg starting fleet cattle io kind cluster controller time level info msg starting gitjob cattle io kind gitjob controller time level info msg starting rbac authorization io kind role controller time level info msg starting kind secret controller time level info msg deployed new agent for cluster fleet local local time level info msg deleted old agent for cluster fleet local local time level info msg deployed new agent for cluster fleet local local time level info msg cluster registration fleet local request hgtpt cluster fleet local local granted time level info msg cluster registration fleet local request hgtpt cluster fleet local local granted time level info msg cluster registration fleet local request hgtpt cluster fleet local local granted time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get context deadline exceeded requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get context deadline exceeded requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get dial tcp i o timeout requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level info msg deleting expired registration secret fleet clusters system c time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c dlkxj handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get context deadline exceeded client timeout exceeded while awaiting headers requeuing time level error msg error syncing fleet default c handler import cluster get net http request canceled while waiting for connection client timeout exceeded while awaiting headers requeuing | 0 |
59,610 | 17,023,177,570 | IssuesEvent | 2021-07-03 00:43:31 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | API returns incorrect Content-Length for elements that have 'gone' | Component: api Priority: trivial Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 10.57am, Friday, 14th September 2007]**
API returns incorrect Content-Length for elements that have 'gone'
For example
If you request http://www.openstreetmap.org/api/0.4/way/4627408
The server returns
Content-Length: 1
Yet no content is returned, so it should return
Content-Length: 0
(or some content) | 1.0 | API returns incorrect Content-Length for elements that have 'gone' - **[Submitted to the original trac issue database at 10.57am, Friday, 14th September 2007]**
API returns incorrect Content-Length for elements that have 'gone'
For example
If you request http://www.openstreetmap.org/api/0.4/way/4627408
The server returns
Content-Length: 1
Yet no content is returned, so it should return
Content-Length: 0
(or some content) | defect | api returns incorrect content length for elements that have gone api returns incorrect content length for elements that have gone for example if you request the server returns content length yet no content is returned so it should return content length or some content | 1 |
63,591 | 17,781,648,615 | IssuesEvent | 2021-08-31 05:46:59 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Created DM doesn't show up under spaces where the target is a member of | T-Defect | ### Steps to reproduce
1. Have a space with someone else as a member
2. DM that person
### What happened?
the DM doesn't show up under the space. Might be a subset of https://github.com/vector-im/element-web/issues/18837
### What did you expect?
for the DM to show up under the space
### Operating system
Windows 10
### Application version
Nightly (2021-08-30)
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Have you submitted a rageshake?
No | 1.0 | Created DM doesn't show up under spaces where the target is a member of - ### Steps to reproduce
1. Have a space with someone else as a member
2. DM that person
### What happened?
the DM doesn't show up under the space. Might be a subset of https://github.com/vector-im/element-web/issues/18837
### What did you expect?
for the DM to show up under the space
### Operating system
Windows 10
### Application version
Nightly (2021-08-30)
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Have you submitted a rageshake?
No | defect | created dm doesn t show up under spaces where the target is a member of steps to reproduce have a space with someone else as a member dm that person what happened the dm doesn t show up under the space might be a subset of what did you expect for the dm to show up under the space operating system windows application version nightly how did you install the app the internet homeserver io have you submitted a rageshake no | 1 |
87,544 | 8,099,089,021 | IssuesEvent | 2018-08-11 03:15:01 | yuzoolcode/pulsecms-tracking | https://api.github.com/repos/yuzoolcode/pulsecms-tracking | closed | JSON code appearing after save | bug needs testing | > JSON_START{"html":"
> Copyright Cake Spirit. 2018 | Privacy policy here <\/a><\/p>\r\n
>
> <\/figure>","meta_searchable":"yes","title":""}JSON_END
> Copyright Cake Spirit. 2018 | Privacy policy here
>
> So I went and removed the image to see what that would do, and code is still there:
>
> JSON_START{"html":"
> Copyright Cake Spirit. 2018 | Privacy policy here <\/a><\/p>\r\n
>
> <\/figure>","meta_searchable":"yes","title":""}JSON_END
> Copyright Cake Spirit. 2018 | Privacy policy here
https://forum.pulsecms.com/t/loads-of-json-code-apeared-on-page-after-editing-a-block/1093 | 1.0 | JSON code appearing after save - > JSON_START{"html":"
> Copyright Cake Spirit. 2018 | Privacy policy here <\/a><\/p>\r\n
>
> <\/figure>","meta_searchable":"yes","title":""}JSON_END
> Copyright Cake Spirit. 2018 | Privacy policy here
>
> So I went and removed the image to see what that would do, and code is still there:
>
> JSON_START{"html":"
> Copyright Cake Spirit. 2018 | Privacy policy here <\/a><\/p>\r\n
>
> <\/figure>","meta_searchable":"yes","title":""}JSON_END
> Copyright Cake Spirit. 2018 | Privacy policy here
https://forum.pulsecms.com/t/loads-of-json-code-apeared-on-page-after-editing-a-block/1093 | non_defect | json code appearing after save json start html copyright cake spirit privacy policy here r n meta searchable yes title json end copyright cake spirit privacy policy here so i went and removed the image to see what that would do and code is still there json start html copyright cake spirit privacy policy here r n meta searchable yes title json end copyright cake spirit privacy policy here | 0 |
4,932 | 2,610,161,534 | IssuesEvent | 2015-02-26 18:51:14 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | ATRT | auto-migrated Priority-Medium Type-Defect | ```
Wrong Tech should be tech 1
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 17 Feb 2011 at 1:10 | 1.0 | ATRT - ```
Wrong Tech should be tech 1
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 17 Feb 2011 at 1:10 | defect | atrt wrong tech should be tech original issue reported on code google com by gmail com on feb at | 1 |
140,269 | 18,901,166,419 | IssuesEvent | 2021-11-16 01:10:57 | KDWSS/aircall-everywhere | https://api.github.com/repos/KDWSS/aircall-everywhere | opened | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz | security vulnerability | ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: aircall-everywhere/package.json</p>
<p>Path to vulnerable library: aircall-everywhere/node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.14.1.tgz (Root Library)
- request-2.88.0.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"node-sass:4.14.1;request:2.88.0;http-signature:1.2.0;jsprim:1.4.1;json-schema:0.2.3","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: aircall-everywhere/package.json</p>
<p>Path to vulnerable library: aircall-everywhere/node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.14.1.tgz (Root Library)
- request-2.88.0.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"node-sass:4.14.1;request:2.88.0;http-signature:1.2.0;jsprim:1.4.1;json-schema:0.2.3","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file aircall everywhere package json path to vulnerable library aircall everywhere node modules json schema package json dependency hierarchy node sass tgz root library request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree node sass request http signature jsprim json schema isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails json schema is vulnerable to improperly controlled modification of object prototype attributes pollution vulnerabilityurl | 0 |
24,967 | 4,157,678,662 | IssuesEvent | 2016-06-16 22:11:22 | opencaching/opencaching-pl | https://api.github.com/repos/opencaching/opencaching-pl | closed | Problem with map in geopath module, when english language is selected | Component_GeoPaths Component_Map Priority_Low Type_Defect | When pl (or any other language is selected) in oc.pl:

When english:

| 1.0 | Problem with map in geopath module, when english language is selected - When pl (or any other language is selected) in oc.pl:

When english:

| defect | problem with map in geopath module when english language is selected when pl or any other language is selected in oc pl when english | 1 |
43,768 | 11,839,084,987 | IssuesEvent | 2020-03-23 16:36:27 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Excess newline in formatted SELECT without FROM clause | C: Functionality E: All Editions P: Low R: Fixed T: Defect | There's an excess newline generated in the formatted rendering of:
``` java
create().with("t", f -> "field_" + f.getName()).as(select(val(1, Integer.class).as("a"), val("a").as("b")))
.select()
.from(table(name("t")))
.fetch();
```
Which produces:
``` sql
with "t"("field_a", "field_b") as (
select
1 as "a",
'a' as "b"
)
select *
from "t"
```
| 1.0 | Excess newline in formatted SELECT without FROM clause - There's an excess newline generated in the formatted rendering of:
``` java
create().with("t", f -> "field_" + f.getName()).as(select(val(1, Integer.class).as("a"), val("a").as("b")))
.select()
.from(table(name("t")))
.fetch();
```
Which produces:
``` sql
with "t"("field_a", "field_b") as (
select
1 as "a",
'a' as "b"
)
select *
from "t"
```
| defect | excess newline in formatted select without from clause there s an excess newline generated in the formatted rendering of java create with t f field f getname as select val integer class as a val a as b select from table name t fetch which produces sql with t field a field b as select as a a as b select from t | 1 |
262,300 | 22,830,507,194 | IssuesEvent | 2022-07-12 12:32:50 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | Update /pages/TestCategoryPage.js to work with future dependencies | component: testing qa: not needed priority: p3 | Update /pages/TestCategoryPage.js to work with future dependencies | 1.0 | Update /pages/TestCategoryPage.js to work with future dependencies - Update /pages/TestCategoryPage.js to work with future dependencies | non_defect | update pages testcategorypage js to work with future dependencies update pages testcategorypage js to work with future dependencies | 0 |
31,610 | 6,558,349,427 | IssuesEvent | 2017-09-06 21:01:01 | westlan-uk/lan-party-site | https://api.github.com/repos/westlan-uk/lan-party-site | closed | Seating plan | auto-migrated Priority-Medium Type-Defect | ```
The new seating plan works well until you come to sign in at the lan and it
asks you for your seat number and shows the old seating plan map :D Fixy fixy :D
An Idea for the lan photo's, being able to tag people in the photos with there
profile name, so if we have new people who get chatting to someone and then
think after the lan they want to talk some more but cant remember the persons
name they can look through the photos and find who they chatted to and get a
tag to their user profile.
Just an idea, make it happen :D only joking mate, just an idea to add to the
long list I am sure you have :D
```
Original issue reported on code.google.com by `xconspir...@gmail.com` on 6 Nov 2012 at 11:58
| 1.0 | Seating plan - ```
The new seating plan works well until you come to sign in at the lan and it
asks you for your seat number and shows the old seating plan map :D Fixy fixy :D
An Idea for the lan photo's, being able to tag people in the photos with there
profile name, so if we have new people who get chatting to someone and then
think after the lan they want to talk some more but cant remember the persons
name they can look through the photos and find who they chatted to and get a
tag to their user profile.
Just an idea, make it happen :D only joking mate, just an idea to add to the
long list I am sure you have :D
```
Original issue reported on code.google.com by `xconspir...@gmail.com` on 6 Nov 2012 at 11:58
| defect | seating plan the new seating plan works well until you come to sign in at the lan and it asks you for your seat number and shows the old seating plan map d fixy fixy d an idea for the lan photo s being able to tag people in the photos with there profile name so if we have new people who get chatting to someone and then think after the lan they want to talk some more but cant remember the persons name they can look through the photos and find who they chatted to and get a tag to their user profile just an idea make it happen d only joking mate just an idea to add to the long list i am sure you have d original issue reported on code google com by xconspir gmail com on nov at | 1 |
779,737 | 27,364,476,790 | IssuesEvent | 2023-02-27 18:04:57 | bbortt/event-planner | https://api.github.com/repos/bbortt/event-planner | closed | fix the problems within ci/cd | bug priority:high ci/cd github_actions | - [ ] the publish job fails because of the invalid jdk11 base image: https://github.com/bbortt/event-planner/actions/runs/4279576918/jobs/7450428894.
- [ ] fix the "unchanged files with check annotations": https://github.com/bbortt/event-planner/pull/668/files | 1.0 | fix the problems within ci/cd - - [ ] the publish job fails because of the invalid jdk11 base image: https://github.com/bbortt/event-planner/actions/runs/4279576918/jobs/7450428894.
- [ ] fix the "unchanged files with check annotations": https://github.com/bbortt/event-planner/pull/668/files | non_defect | fix the problems within ci cd the publish job fails because of the invalid base image fix the unchanged files with check annotations | 0 |
676,321 | 23,122,280,889 | IssuesEvent | 2022-07-27 23:17:29 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | Implement SCHECK alternative that will behave like CHECK in debug | kind/enhancement area/docdb priority/medium | Jira Link: [DB-1641](https://yugabyte.atlassian.net/browse/DB-1641)
This should have the following behavior depending on build type:
- Release: Return status if the condition is not met (like `SCHECK`)
- Debug: Crash if the condition is not met (like `CHECK`)
| 1.0 | Implement SCHECK alternative that will behave like CHECK in debug - Jira Link: [DB-1641](https://yugabyte.atlassian.net/browse/DB-1641)
This should have the following behavior depending on build type:
- Release: Return status if the condition is not met (like `SCHECK`)
- Debug: Crash if the condition is not met (like `CHECK`)
| non_defect | implement scheck alternative that will behave like check in debug jira link this should have the following behavior depending on build type release return status if the condition is not met like scheck debug crash if the condition is not met like check | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.