Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,190 | 19,211,511,002 | IssuesEvent | 2021-12-07 02:50:49 | actiontech/sqle | https://api.github.com/repos/actiontech/sqle | closed | support multi-params configuration of rule | ui feature not_compatible | ### 背景
1. #55 中需要新增 3 个配置,层级不同,无法通过 3 条规则友好的表示
2. 当前规则只支持配置 1 个规则值,无法支持复杂规则的自定义
3. 当前规则值以字符串的形式展示,用户在修改规则值时,缺少类型限制(用户可能在规则值中填写任何值) | True | support multi-params configuration of rule - ### 背景
1. #55 中需要新增 3 个配置,层级不同,无法通过 3 条规则友好的表示
2. 当前规则只支持配置 1 个规则值,无法支持复杂规则的自定义
3. 当前规则值以字符串的形式展示,用户在修改规则值时,缺少类型限制(用户可能在规则值中填写任何值) | non_main | support multi params configuration of rule 背景 中需要新增 个配置,层级不同,无法通过 条规则友好的表示 当前规则只支持配置 个规则值,无法支持复杂规则的自定义 当前规则值以字符串的形式展示,用户在修改规则值时,缺少类型限制(用户可能在规则值中填写任何值) | 0 |
5,402 | 27,115,680,630 | IssuesEvent | 2023-02-15 18:22:29 | VA-Explorer/va_explorer | https://api.github.com/repos/VA-Explorer/va_explorer | closed | Display 'mean age of death' info to demographic tab | Type: Maintainance Language: Python Domain: Frontend Status: Inactive | **What is the expected state?**
I expect to be able to view 'mean age of death' for each demographic highlighted on the dashboard demographic tab. Perhaps via the already existing hover tooltip.
**What is the actual state?**
Currently highlighted demographics of gender, age group, and location of death display the percentage of the total + the exact count.
**Relevant context**
- We already display mean age of death on the map location tooltips (broken down by location), this change would adapt that to also display, broken down by demographic
- `va_analytics/dash_apps/va_dashboard.py`
- `va_analytics/utils/plotting.py`
- `va_data_management/utils/validate.py` | True | Display 'mean age of death' info to demographic tab - **What is the expected state?**
I expect to be able to view 'mean age of death' for each demographic highlighted on the dashboard demographic tab. Perhaps via the already existing hover tooltip.
**What is the actual state?**
Currently highlighted demographics of gender, age group, and location of death display the percentage of the total + the exact count.
**Relevant context**
- We already display mean age of death on the map location tooltips (broken down by location), this change would adapt that to also display, broken down by demographic
- `va_analytics/dash_apps/va_dashboard.py`
- `va_analytics/utils/plotting.py`
- `va_data_management/utils/validate.py` | main | display mean age of death info to demographic tab what is the expected state i expect to be able to view mean age of death for each demographic highlighted on the dashboard demographic tab perhaps via the already existing hover tooltip what is the actual state currently highlighted demographics of gender age group and location of death display the percentage of the total the exact count relevant context we already display mean age of death on the map location tooltips broken down by location this change would adapt that to also display broken down by demographic va analytics dash apps va dashboard py va analytics utils plotting py va data management utils validate py | 1 |
709,552 | 24,382,723,364 | IssuesEvent | 2022-10-04 09:11:26 | tum-ei-eda/mlonmcu | https://api.github.com/repos/tum-ei-eda/mlonmcu | closed | Extend Platform API: BuildPlatform | priority:low | A platform API was recently added to MLonMCU and it's idea can be described as follows:
- Common interface: `Platform`
- Specific platform types/classes:
- `CompilePlatform`: A platform which is able to build target software with given codegen results
- `TargetPlatform`: A platform with the ability to flash/monitor specific (hardware) targets
- A certain platform (say `espidf` or `platformio`) inherits from one or both of the base classes depending on the implemented features
- Platforms can also be seen as a "target registry" as they are able to dynamically create `Target` instances for supported target names.
This Issue proposes to add another type of platforms: `BuildPlatform`
This would be a platform which wraps around a backend and should be therefore able to run code generation.
A realistic example of how this might be use would be a `microtvm` platform as TVM provides a `Project API` with templates to support large number of target devices. The full flow from building a model over compiling the running the model (using an RPC server) can be handled using the `tvmc micro` tool.
I am actually not sure if this would be a good idea at some point in time, however it always make sense to think about ways to generalize existing APIs. For this reason this should mainly be documenting the concept which might get picked up at some point. | 1.0 | Extend Platform API: BuildPlatform - A platform API was recently added to MLonMCU and it's idea can be described as follows:
- Common interface: `Platform`
- Specific platform types/classes:
- `CompilePlatform`: A platform which is able to build target software with given codegen results
- `TargetPlatform`: A platform with the ability to flash/monitor specific (hardware) targets
- A certain platform (say `espidf` or `platformio`) inherits from one or both of the base classes depending on the implemented features
- Platforms can also be seen as a "target registry" as they are able to dynamically create `Target` instances for supported target names.
This Issue proposes to add another type of platforms: `BuildPlatform`
This would be a platform which wraps around a backend and should be therefore able to run code generation.
A realistic example of how this might be use would be a `microtvm` platform as TVM provides a `Project API` with templates to support large number of target devices. The full flow from building a model over compiling the running the model (using an RPC server) can be handled using the `tvmc micro` tool.
I am actually not sure if this would be a good idea at some point in time, however it always make sense to think about ways to generalize existing APIs. For this reason this should mainly be documenting the concept which might get picked up at some point. | non_main | extend platform api buildplatform a platform api was recently added to mlonmcu and it s idea can be described as follows common interface platform specific platform types classes compileplatform a platform which is able to build target software with given codegen results targetplatform a platform with the ability to flash monitor specific hardware targets a certain platform say espidf or platformio inherits from one or both of the base classes depending on the implemented features platforms can also be seen as a target registry as they are able to dynamically create target instances for supported target names this issue proposes to add another type of platforms buildplatform this would be a platform which wraps around a backend and should be therefore able to run code generation a realistic example of how this might be use would be a microtvm platform as tvm provides a project api with templates to support large number of target devices the full flow from building a model over compiling the running the model using an rpc server can be handled using the tvmc micro tool i am actually not sure if this would be a good idea at some point in time however it always make sense to think about ways to generalize existing apis for this reason this should mainly be documenting the concept which might get picked up at some point | 0 |
17,058 | 2,615,129,917 | IssuesEvent | 2015-03-01 05:59:42 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | Release Subscription Notifications support | auto-migrated Milestone-Version1.14.0 Priority-High Type-Enhancement | ```
Google's subscriptions protocol is changing, and we'll use this issue to track
those changes. These changes are final yet, so I'll wait to later to fill in
those details.
For now I'll just mention that I'd like to move Notification.EVENT_TYPE_HEADER
to AbstractWebHookServlet.
```
Original issue reported on code.google.com by `yan...@google.com` on 1 Nov 2012 at 4:56 | 1.0 | Release Subscription Notifications support - ```
Google's subscriptions protocol is changing, and we'll use this issue to track
those changes. These changes are final yet, so I'll wait to later to fill in
those details.
For now I'll just mention that I'd like to move Notification.EVENT_TYPE_HEADER
to AbstractWebHookServlet.
```
Original issue reported on code.google.com by `yan...@google.com` on 1 Nov 2012 at 4:56 | non_main | release subscription notifications support google s subscriptions protocol is changing and we ll use this issue to track those changes these changes are final yet so i ll wait to later to fill in those details for now i ll just mention that i d like to move notification event type header to abstractwebhookservlet original issue reported on code google com by yan google com on nov at | 0 |
715,477 | 24,600,751,837 | IssuesEvent | 2022-10-14 12:18:30 | ballerina-platform/ballerina-standard-library | https://api.github.com/repos/ballerina-platform/ballerina-standard-library | closed | [Asgardeo to Choreo Hub] Cannot run multiple instances of `hub` | Points/4 Priority/Blocker Type/Bug Team/PCM | **Description:**
When running multiple `hub` instances duplicate event notifications will be send to the subscribers.
| 1.0 | [Asgardeo to Choreo Hub] Cannot run multiple instances of `hub` - **Description:**
When running multiple `hub` instances duplicate event notifications will be send to the subscribers.
| non_main | cannot run multiple instances of hub description when running multiple hub instances duplicate event notifications will be send to the subscribers | 0 |
59,700 | 8,376,996,575 | IssuesEvent | 2018-10-05 22:02:25 | opal/opal | https://api.github.com/repos/opal/opal | closed | Browser support | discuss documentation | > On an unrelated subject (well, maybe not so unrelated), I believe the claim that Opal runs on IE 6 is no longer true. I have seen things like Object.keys used in corelib (I think in the Array or Hash implementation), and that only became available in IE 9 I think. Not sure what if anything needs to be done about that, but wanted to hear your thoughts on what browsers should (ideally vs practically) be supported and how to make sure they are as work on Opal continues.
> — @vais in #726
cc @adambeynon @meh
| 1.0 | Browser support - > On an unrelated subject (well, maybe not so unrelated), I believe the claim that Opal runs on IE 6 is no longer true. I have seen things like Object.keys used in corelib (I think in the Array or Hash implementation), and that only became available in IE 9 I think. Not sure what if anything needs to be done about that, but wanted to hear your thoughts on what browsers should (ideally vs practically) be supported and how to make sure they are as work on Opal continues.
> — @vais in #726
cc @adambeynon @meh
| non_main | browser support on an unrelated subject well maybe not so unrelated i believe the claim that opal runs on ie is no longer true i have seen things like object keys used in corelib i think in the array or hash implementation and that only became available in ie i think not sure what if anything needs to be done about that but wanted to hear your thoughts on what browsers should ideally vs practically be supported and how to make sure they are as work on opal continues — vais in cc adambeynon meh | 0 |
728 | 4,318,963,280 | IssuesEvent | 2016-07-24 11:09:25 | gogits/gogs | https://api.github.com/repos/gogits/gogs | closed | Pull request gives 500 error after deleting "base" branch from old pull request | kind/bug status/assigned to maintainer status/needs feedback | - Gogs version (or commit ref): 0.9.13.0319
- Git version: 2.8.1.windows.1
- Operating system: Windows Server 2012 R2
- Database:
- [ ] PostgreSQL
- [ ] MySQL
- [ ] SQLite
- Can you reproduce the bug at http://try.gogs.io: https://try.gogs.io/apettit/test500error/
- [X ] Yes (provide example URL)
- [ ] No
- [ ] Not relevant
- Log gist:[E] NewPullRequest: testPatch: UpdateLocalCopy: Pull: exit status 1 - Your configuration specifies to merge with the ref 'refs/heads/branch_to_delete'
## Description
You will get a 500 error if you delete a "base" branch of an old pull request.
Reproduction steps
1. Start with one branch "master" with a "test.txt" file
2. Make 2 additional branches "test" and "branch_to_delete" from master and push them
3. Change the contents of test.txtC
4. heck in on "test" branch and push
5. In GOGs create a pull request to merge the branch "test" into "branch_to_delete". Do not accept or merge pull request at this time.
6. Delete branch "branch_to_delete" (origin/branch_to_delete)
7. In GOGs attempt to create a pull request to merge "test" into "master". The initial diff will work, but when you click "create pull request" you will get a 500 error.
Additional info: You also get a 500 error if you try to back into the first pull request to close it. | True | Pull request gives 500 error after deleting "base" branch from old pull request - - Gogs version (or commit ref): 0.9.13.0319
- Git version: 2.8.1.windows.1
- Operating system: Windows Server 2012 R2
- Database:
- [ ] PostgreSQL
- [ ] MySQL
- [ ] SQLite
- Can you reproduce the bug at http://try.gogs.io: https://try.gogs.io/apettit/test500error/
- [X ] Yes (provide example URL)
- [ ] No
- [ ] Not relevant
- Log gist:[E] NewPullRequest: testPatch: UpdateLocalCopy: Pull: exit status 1 - Your configuration specifies to merge with the ref 'refs/heads/branch_to_delete'
## Description
You will get a 500 error if you delete a "base" branch of an old pull request.
Reproduction steps
1. Start with one branch "master" with a "test.txt" file
2. Make 2 additional branches "test" and "branch_to_delete" from master and push them
3. Change the contents of test.txtC
4. heck in on "test" branch and push
5. In GOGs create a pull request to merge the branch "test" into "branch_to_delete". Do not accept or merge pull request at this time.
6. Delete branch "branch_to_delete" (origin/branch_to_delete)
7. In GOGs attempt to create a pull request to merge "test" into "master". The initial diff will work, but when you click "create pull request" you will get a 500 error.
Additional info: You also get a 500 error if you try to back into the first pull request to close it. | main | pull request gives error after deleting base branch from old pull request gogs version or commit ref git version windows operating system windows server database postgresql mysql sqlite can you reproduce the bug at yes provide example url no not relevant log gist newpullrequest testpatch updatelocalcopy pull exit status your configuration specifies to merge with the ref refs heads branch to delete description you will get a error if you delete a base branch of an old pull request reproduction steps start with one branch master with a test txt file make additional branches test and branch to delete from master and push them change the contents of test txtc heck in on test branch and push in gogs create a pull request to merge the branch test into branch to delete do not accept or merge pull request at this time delete branch branch to delete origin branch to delete in gogs attempt to create a pull request to merge test into master the initial diff will work but when you click create pull request you will get a error additional info you also get a error if you try to back into the first pull request to close it | 1 |
13,538 | 23,297,052,203 | IssuesEvent | 2022-08-06 18:49:00 | kysect/Shreks | https://api.github.com/repos/kysect/Shreks | closed | Фейковый API для эмулирования активности на гитхабе | requirement blocked area:github | Для разработки и тестирования нужно будет очень много действий делать на гитхабе и будет сильно проще, если будет возможность эмулировать действия на гитхабе. Пока это кажется как какой-то REST API, который умеет в те же действия, которые мы хендлин от вебхуков.
## Blocked by:
- #1 | 1.0 | Фейковый API для эмулирования активности на гитхабе - Для разработки и тестирования нужно будет очень много действий делать на гитхабе и будет сильно проще, если будет возможность эмулировать действия на гитхабе. Пока это кажется как какой-то REST API, который умеет в те же действия, которые мы хендлин от вебхуков.
## Blocked by:
- #1 | non_main | фейковый api для эмулирования активности на гитхабе для разработки и тестирования нужно будет очень много действий делать на гитхабе и будет сильно проще если будет возможность эмулировать действия на гитхабе пока это кажется как какой то rest api который умеет в те же действия которые мы хендлин от вебхуков blocked by | 0 |
19,442 | 26,992,522,411 | IssuesEvent | 2023-02-09 21:11:30 | VazkiiMods/Botania | https://api.github.com/repos/VazkiiMods/Botania | closed | Implement Fabric transfer API support for Corporea | compatibility confirmed | ### Mod Loader
Fabric
### Minecraft Version
1.18.2
### Botania version
1.18.2-434-FABRIC
### Modloader version
Loader: 0.14.9 Api: 0.58.0+1.18.2
### Modpack info
Extended Drawers 1.3.3+mc1.18.2
### The latest.log file
https://pastebin.com/eJ20S58M
### Issue description
Corporea sparks can't be attached to blocks from extended drawers. I suspect it's because I exclusively use the fabric transfer api, but I can't confirm it as the corporea code really confuses me.
### Steps to reproduce
1. Set up a corporea network
2. Place down a drawer
3. Fail trying to connect the drawers
### Other information
Originally reported here: MattiDragon/extendeddrawers#18 | True | Implement Fabric transfer API support for Corporea - ### Mod Loader
Fabric
### Minecraft Version
1.18.2
### Botania version
1.18.2-434-FABRIC
### Modloader version
Loader: 0.14.9 Api: 0.58.0+1.18.2
### Modpack info
Extended Drawers 1.3.3+mc1.18.2
### The latest.log file
https://pastebin.com/eJ20S58M
### Issue description
Corporea sparks can't be attached to blocks from extended drawers. I suspect it's because I exclusively use the fabric transfer api, but I can't confirm it as the corporea code really confuses me.
### Steps to reproduce
1. Set up a corporea network
2. Place down a drawer
3. Fail trying to connect the drawers
### Other information
Originally reported here: MattiDragon/extendeddrawers#18 | non_main | implement fabric transfer api support for corporea mod loader fabric minecraft version botania version fabric modloader version loader api modpack info extended drawers the latest log file issue description corporea sparks can t be attached to blocks from extended drawers i suspect it s because i exclusively use the fabric transfer api but i can t confirm it as the corporea code really confuses me steps to reproduce set up a corporea network place down a drawer fail trying to connect the drawers other information originally reported here mattidragon extendeddrawers | 0 |
2,878 | 8,461,748,442 | IssuesEvent | 2018-10-22 23:01:16 | poanetwork/blockscout | https://api.github.com/repos/poanetwork/blockscout | opened | Geth Clique blocks always mined by 0x000 | enhancement priority: high team: architecture | In order to find the miner for Geth Clique blocks, we must use `clique.getSignersAtHash('0x86837f148f096f1ce53463ecb4acf3d99bae0b2f0297dacbb6d33fcfdb7c04d9')`.
| 1.0 | Geth Clique blocks always mined by 0x000 - In order to find the miner for Geth Clique blocks, we must use `clique.getSignersAtHash('0x86837f148f096f1ce53463ecb4acf3d99bae0b2f0297dacbb6d33fcfdb7c04d9')`.
| non_main | geth clique blocks always mined by in order to find the miner for geth clique blocks we must use clique getsignersathash | 0 |
73,808 | 14,118,792,801 | IssuesEvent | 2020-11-08 15:07:41 | rtcharity/eahub.org | https://api.github.com/repos/rtcharity/eahub.org | closed | Use Django forms for rendering | Code Quality Velocity | See, e.g., https://docs.djangoproject.com/en/2.1/topics/forms/
This will make it possible to change the workings of models without writing a bunch more HTML.
django-crispy-forms is probably the easiest way to integrate this with Bootstrap. | 1.0 | Use Django forms for rendering - See, e.g., https://docs.djangoproject.com/en/2.1/topics/forms/
This will make it possible to change the workings of models without writing a bunch more HTML.
django-crispy-forms is probably the easiest way to integrate this with Bootstrap. | non_main | use django forms for rendering see e g this will make it possible to change the workings of models without writing a bunch more html django crispy forms is probably the easiest way to integrate this with bootstrap | 0 |
201,308 | 22,948,166,468 | IssuesEvent | 2022-07-19 03:39:51 | elikkatzgit/TestingPOM | https://api.github.com/repos/elikkatzgit/TestingPOM | reopened | CVE-2021-20190 (High) detected in jackson-databind-2.7.2.jar | security vulnerability | ## CVE-2021-20190 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in jackson-databind before 2.9.10.7. FasterXML mishandles the interaction between serialization gadgets and typing. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability.
<p>Publish Date: 2021-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20190>CVE-2021-20190</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-19</p>
<p>Fix Resolution: 2.9.10.7</p>
</p>
</details>
<p></p>
| True | CVE-2021-20190 (High) detected in jackson-databind-2.7.2.jar - ## CVE-2021-20190 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in jackson-databind before 2.9.10.7. FasterXML mishandles the interaction between serialization gadgets and typing. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability.
<p>Publish Date: 2021-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20190>CVE-2021-20190</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-19</p>
<p>Fix Resolution: 2.9.10.7</p>
</p>
</details>
<p></p>
| non_main | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href dependency hierarchy x jackson databind jar vulnerable library found in base branch dev vulnerability details a flaw was found in jackson databind before fasterxml mishandles the interaction between serialization gadgets and typing the highest threat from this vulnerability is to data confidentiality and integrity as well as system availability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution | 0 |
29,891 | 4,540,426,235 | IssuesEvent | 2016-09-09 14:37:12 | researchstudio-sat/webofneeds | https://api.github.com/repos/researchstudio-sat/webofneeds | closed | MobileFirst Changes | enhancement testing UX | Check/Review and Change the current GUI so that it works and looks nice on a mobile device as well | 1.0 | MobileFirst Changes - Check/Review and Change the current GUI so that it works and looks nice on a mobile device as well | non_main | mobilefirst changes check review and change the current gui so that it works and looks nice on a mobile device as well | 0 |
167,735 | 13,040,839,021 | IssuesEvent | 2020-07-28 19:14:52 | apple/servicetalk | https://api.github.com/repos/apple/servicetalk | closed | Flaky test: DefaultDnsClientTest.repeatDiscoverNxDomain | flaky tests | https://ci.servicetalk.io/job/servicetalk-java8-prb/1450/testReport/junit/io.servicetalk.dns.discovery.netty/DefaultDnsClientTest/repeatDiscoverNxDomain/
```
Regression
io.servicetalk.dns.discovery.netty.DefaultDnsClientTest.repeatDiscoverNxDomain
Error Message
java.lang.AssertionError: Unexpected exception during DNS lookup.
Expected: an instance of java.net.UnknownHostException
but: null
Stacktrace
java.lang.AssertionError: Unexpected exception during DNS lookup.
Expected: an instance of java.net.UnknownHostException
but: null
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at io.servicetalk.dns.discovery.netty.DefaultDnsClientTest.repeatDiscoverNxDomain(DefaultDnsClientTest.java:478)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:171)
at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:163)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
``` | 1.0 | Flaky test: DefaultDnsClientTest.repeatDiscoverNxDomain - https://ci.servicetalk.io/job/servicetalk-java8-prb/1450/testReport/junit/io.servicetalk.dns.discovery.netty/DefaultDnsClientTest/repeatDiscoverNxDomain/
```
Regression
io.servicetalk.dns.discovery.netty.DefaultDnsClientTest.repeatDiscoverNxDomain
Error Message
java.lang.AssertionError: Unexpected exception during DNS lookup.
Expected: an instance of java.net.UnknownHostException
but: null
Stacktrace
java.lang.AssertionError: Unexpected exception during DNS lookup.
Expected: an instance of java.net.UnknownHostException
but: null
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at io.servicetalk.dns.discovery.netty.DefaultDnsClientTest.repeatDiscoverNxDomain(DefaultDnsClientTest.java:478)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:171)
at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:163)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
``` | non_main | flaky test defaultdnsclienttest repeatdiscovernxdomain regression io servicetalk dns discovery netty defaultdnsclienttest repeatdiscovernxdomain error message java lang assertionerror unexpected exception during dns lookup expected an instance of java net unknownhostexception but null stacktrace java lang assertionerror unexpected exception during dns lookup expected an instance of java net unknownhostexception but null at org hamcrest matcherassert assertthat matcherassert java at io servicetalk dns discovery netty defaultdnsclienttest repeatdiscovernxdomain defaultdnsclienttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement callablestatement call servicetalktesttimeout java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement callablestatement call servicetalktesttimeout java at java util concurrent futuretask run futuretask java at java lang thread run thread java | 0 |
441,080 | 12,707,600,379 | IssuesEvent | 2020-06-23 09:10:11 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | retail.onlinesbi.com - see bug description | browser-fenix engine-gecko priority-critical | <!-- @browser: Firefox Mobile 77.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/54354 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://retail.onlinesbi.com/retail/login.htm
**Browser / Version**: Firefox Mobile 77.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Login not visible
**Steps to Reproduce**:
Login not visible on all websites.
Kindly email me the solution.
manavsharma875@gmail.com
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | retail.onlinesbi.com - see bug description - <!-- @browser: Firefox Mobile 77.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/54354 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://retail.onlinesbi.com/retail/login.htm
**Browser / Version**: Firefox Mobile 77.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Login not visible
**Steps to Reproduce**:
Login not visible on all websites.
Kindly email me the solution.
manavsharma875@gmail.com
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | retail onlinesbi com see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description login not visible steps to reproduce login not visible on all websites kindly email me the solution gmail com browser configuration none from with ❤️ | 0 |
3,466 | 13,302,864,538 | IssuesEvent | 2020-08-25 14:47:21 | PowerShell/PowerShell | https://api.github.com/repos/PowerShell/PowerShell | closed | Assemblies listed in FileList Module Manifest Property are being Processed. | Issue-Question Review - Maintainer | <!--
For Windows PowerShell 5.1 issues, suggestions, or feature requests please use the following link instead:
Windows PowerShell [UserVoice](https://windowsserver.uservoice.com/forums/301869-powershell)
This repository is **ONLY** for PowerShell Core 6 and PowerShell 7+ issues.
- Make sure you are able to repro it on the [latest released version](https://github.com/PowerShell/PowerShell/releases)
- Search the existing issues.
- Refer to the [FAQ](https://github.com/PowerShell/PowerShell/blob/master/docs/FAQ.md).
- Refer to the [known issues](https://docs.microsoft.com/powershell/scripting/whats-new/known-issues-ps6).
-->
## Steps to reproduce
I have a WinSCP PowerShell Module Wrapper, and I was implementing some functionality to load one WinSCPnet.dll if it is PowerShell core, and a different one if it is PowerShell desktop. And I kept getting errors stating that the assembly of the same name was already loaded.
Turns out, I had both of the assemblies listed in the Module Manifest `FileList` property, after removing them from there, the logic to add the correct assembly works fine. Per this link, that property should be for inventory only, and should not process the files in the value: https://docs.microsoft.com/en-us/powershell/scripting/developer/module/how-to-write-a-powershell-module-manifest?view=powershell-7#module-manifest-elements
```powershell
# Assembly loading logic from WinSCP.psm1
$moduleRoot = Split-Path -Path $MyInvocation.MyCommand.Path
switch ($PSVersionTable.PSEdition) {
"Core" {
#Add-Type -Path "${moduleRoot}\lib\netstandard2.0\WinSCPnet.dll"
add-type -path ./Documents/github/WinSCP/WinSCP/lib/netstandard2.0/WinSCPnet.dll
break;
}
"Desktop" {
Add-Type -Path "${moduleRoot}\lib\net40\WinSCPnet.dll"
break;
}
default {
Write-Error -Message "Failed to find a compatiable WinSCP Assembly."
exit
}
}
```
When the `FileList` property in the manifest has those files listed in the value, I get the "Assembly with the same name is already loaded" error


if I remove the .dll values from that list, the logic works fine and the proper dll is loaded.
## Expected behavior
No processing of the files in the `FileList` module manifest property value.
## Actual behavior
Im not sure in what way, but he files are being executed or processed.
## Environment data
<!-- provide the output of $PSVersionTable -->
```none
PS /Users/thomas> $psversiontable
Name Value
---- -----
PSVersion 7.0.3
PSEdition Core
GitCommitId 7.0.3
OS Darwin 20.0.0 Darwin Kernel Version 20.0.0: Thu…
Platform Unix
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
```
| True | Assemblies listed in FileList Module Manifest Property are being Processed. - <!--
For Windows PowerShell 5.1 issues, suggestions, or feature requests please use the following link instead:
Windows PowerShell [UserVoice](https://windowsserver.uservoice.com/forums/301869-powershell)
This repository is **ONLY** for PowerShell Core 6 and PowerShell 7+ issues.
- Make sure you are able to repro it on the [latest released version](https://github.com/PowerShell/PowerShell/releases)
- Search the existing issues.
- Refer to the [FAQ](https://github.com/PowerShell/PowerShell/blob/master/docs/FAQ.md).
- Refer to the [known issues](https://docs.microsoft.com/powershell/scripting/whats-new/known-issues-ps6).
-->
## Steps to reproduce
I have a WinSCP PowerShell Module Wrapper, and I was implementing some functionality to load one WinSCPnet.dll if it is PowerShell core, and a different one if it is PowerShell desktop. And I kept getting errors stating that the assembly of the same name was already loaded.
Turns out, I had both of the assemblies listed in the Module Manifest `FileList` property, after removing them from there, the logic to add the correct assembly works fine. Per this link, that property should be for inventory only, and should not process the files in the value: https://docs.microsoft.com/en-us/powershell/scripting/developer/module/how-to-write-a-powershell-module-manifest?view=powershell-7#module-manifest-elements
```powershell
# Assembly loading logic from WinSCP.psm1
$moduleRoot = Split-Path -Path $MyInvocation.MyCommand.Path
switch ($PSVersionTable.PSEdition) {
"Core" {
#Add-Type -Path "${moduleRoot}\lib\netstandard2.0\WinSCPnet.dll"
add-type -path ./Documents/github/WinSCP/WinSCP/lib/netstandard2.0/WinSCPnet.dll
break;
}
"Desktop" {
Add-Type -Path "${moduleRoot}\lib\net40\WinSCPnet.dll"
break;
}
default {
Write-Error -Message "Failed to find a compatiable WinSCP Assembly."
exit
}
}
```
When the `FileList` property in the manifest has those files listed in the value, I get the "Assembly with the same name is already loaded" error


if I remove the .dll values from that list, the logic works fine and the proper dll is loaded.
## Expected behavior
No processing of the files in the `FileList` module manifest property value.
## Actual behavior
Im not sure in what way, but he files are being executed or processed.
## Environment data
<!-- provide the output of $PSVersionTable -->
```none
PS /Users/thomas> $psversiontable
Name Value
---- -----
PSVersion 7.0.3
PSEdition Core
GitCommitId 7.0.3
OS Darwin 20.0.0 Darwin Kernel Version 20.0.0: Thu…
Platform Unix
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
```
| main | assemblies listed in filelist module manifest property are being processed for windows powershell issues suggestions or feature requests please use the following link instead windows powershell this repository is only for powershell core and powershell issues make sure you are able to repro it on the search the existing issues refer to the refer to the steps to reproduce i have a winscp powershell module wrapper and i was implementing some functionality to load one winscpnet dll if it is powershell core and a different one if it is powershell desktop and i kept getting errors stating that the assembly of the same name was already loaded turns out i had both of the assemblies listed in the module manifest filelist property after removing them from there the logic to add the correct assembly works fine per this link that property should be for inventory only and should not process the files in the value powershell assembly loading logic from winscp moduleroot split path path myinvocation mycommand path switch psversiontable psedition core add type path moduleroot lib winscpnet dll add type path documents github winscp winscp lib winscpnet dll break desktop add type path moduleroot lib winscpnet dll break default write error message failed to find a compatiable winscp assembly exit when the filelist property in the manifest has those files listed in the value i get the assembly with the same name is already loaded error if i remove the dll values from that list the logic works fine and the proper dll is loaded expected behavior no processing of the files in the filelist module manifest property value actual behavior im not sure in what way but he files are being executed or processed environment data none ps users thomas psversiontable name value psversion psedition core gitcommitid os darwin darwin kernel version thu… platform unix pscompatibleversions … psremotingprotocolversion serializationversion wsmanstackversion | 1 |
690,681 | 23,669,184,586 | IssuesEvent | 2022-08-27 04:26:14 | ecotiya/ecotiya-portfolio-site | https://api.github.com/repos/ecotiya/ecotiya-portfolio-site | opened | AWS利用料金の全面見直し | bug Priority: Critical | <!-- 要望のテンプレート -->
## 概要
AWS利用料金の全面見直し
## 目的
あまりにも高すぎるので、不要なところを削る。そのためには、購入した本の見直しが必要。
マルチAZ構成である必要性がおそらくありません。
## タスク
- [ ] [AWSコンテナ設計・構築[本格]入門](https://www.amazon.co.jp/gp/product/B09DKZC1ZH/ref=ppx_yo_dt_b_search_asin_title?ie=UTF8&psc=1)を参照し、ネットワーク構成の見直しを実施する。
- [ ] 現行マルチAZ構成になっているため、シングルAZに切り替えするよう、[AWS CloudFormation](https://aws.amazon.com/jp/cloudformation/)のコードを書き換えする。
- [ ] 環境の再設定を実施し、ソースコードを再度デプロイする。
## 補足
AWS利用料金が高すぎるので、早く改善したい。2022年9月中には片付けておきたい。
全額経費になるとはいえ、このままじゃいかないと思います。 | 1.0 | AWS利用料金の全面見直し - <!-- 要望のテンプレート -->
## 概要
AWS利用料金の全面見直し
## 目的
あまりにも高すぎるので、不要なところを削る。そのためには、購入した本の見直しが必要。
マルチAZ構成である必要性がおそらくありません。
## タスク
- [ ] [AWSコンテナ設計・構築[本格]入門](https://www.amazon.co.jp/gp/product/B09DKZC1ZH/ref=ppx_yo_dt_b_search_asin_title?ie=UTF8&psc=1)を参照し、ネットワーク構成の見直しを実施する。
- [ ] 現行マルチAZ構成になっているため、シングルAZに切り替えするよう、[AWS CloudFormation](https://aws.amazon.com/jp/cloudformation/)のコードを書き換えする。
- [ ] 環境の再設定を実施し、ソースコードを再度デプロイする。
## 補足
AWS利用料金が高すぎるので、早く改善したい。2022年9月中には片付けておきたい。
全額経費になるとはいえ、このままじゃいかないと思います。 | non_main | aws利用料金の全面見直し 概要 aws利用料金の全面見直し 目的 あまりにも高すぎるので、不要なところを削る。そのためには、購入した本の見直しが必要。 マルチaz構成である必要性がおそらくありません。 タスク 現行マルチaz構成になっているため、シングルazに切り替えするよう、 環境の再設定を実施し、ソースコードを再度デプロイする。 補足 aws利用料金が高すぎるので、早く改善したい。 。 全額経費になるとはいえ、このままじゃいかないと思います。 | 0 |
24,099 | 5,028,726,605 | IssuesEvent | 2016-12-15 19:05:07 | julien-truffaut/Monocle | https://api.github.com/repos/julien-truffaut/Monocle | opened | Add documentation about Apply optics | documentation | Idea of content:
- description: partially applied optics (`ApplyLens[S, A] == (S + Lens[S, A])`, `ApplyIso[S, A] == (S, Iso[S, A])`, etc)
- allow to use optics as an infix operator: `s lens set a` instead of `lens.set(a)(s)`
- help type inference and implicit resolution
- example of apply lens syntax `person.lens(_.address.streeName).get` | 1.0 | Add documentation about Apply optics - Idea of content:
- description: partially applied optics (`ApplyLens[S, A] == (S + Lens[S, A])`, `ApplyIso[S, A] == (S, Iso[S, A])`, etc)
- allow to use optics as an infix operator: `s lens set a` instead of `lens.set(a)(s)`
- help type inference and implicit resolution
- example of apply lens syntax `person.lens(_.address.streeName).get` | non_main | add documentation about apply optics idea of content description partially applied optics applylens s lens applyiso s iso etc allow to use optics as an infix operator s lens set a instead of lens set a s help type inference and implicit resolution example of apply lens syntax person lens address streename get | 0 |
68,709 | 8,321,561,229 | IssuesEvent | 2018-09-26 00:52:53 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | opened | Improve datepicker styles | Needs Design | There's been discussion in #7621 around improving the styling/UX of the datepicker. @jasmussen has some great ideas here: https://github.com/WordPress/gutenberg/pull/7621#issuecomment-401301889 and in the same issue @karmatosed referenced a datepicker she dug: https://github.com/WordPress/gutenberg/pull/7621#issuecomment-401605022
We should do these things for sure, but I'm creating this issue separately to document the styles/UX we want to implement that we can work on after #7621 is merged. | 1.0 | Improve datepicker styles - There's been discussion in #7621 around improving the styling/UX of the datepicker. @jasmussen has some great ideas here: https://github.com/WordPress/gutenberg/pull/7621#issuecomment-401301889 and in the same issue @karmatosed referenced a datepicker she dug: https://github.com/WordPress/gutenberg/pull/7621#issuecomment-401605022
We should do these things for sure, but I'm creating this issue separately to document the styles/UX we want to implement that we can work on after #7621 is merged. | non_main | improve datepicker styles there s been discussion in around improving the styling ux of the datepicker jasmussen has some great ideas here and in the same issue karmatosed referenced a datepicker she dug we should do these things for sure but i m creating this issue separately to document the styles ux we want to implement that we can work on after is merged | 0 |
42,784 | 5,535,884,827 | IssuesEvent | 2017-03-21 18:19:22 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | With clause in query expressions (support for zip) | 0 - Backlog Area-Language Design Feature Request | I propose to add a `with` clause to the query expression syntax. This will allow us to zip sequences together using either LINQ syntax.
A query expression with a `with` clause followed by a `select` clause
``` c#
from x1 in e1
with x2 in e2
select v
```
is translated into
``` c#
( e1 ) . Zip( e2, ( x1 , x2 ) => v )
```
A query expression with a `with` clause followed by something other than a `select` clause
``` c#
from x1 in e1
with x2 in e2
…
```
is translated into
``` c#
from * in ( e1 ) . Zip(
e2 , ( x1 , x2 ) => new { x1 , x2 })
…
```
This transformation would happen in section 7.16.2.4 of the spec:
``` c#
from x1 in e1
from x2 in e2
with x3 in e3
…
```
is first translated into
``` c#
from * in ( e1 ) . SelectMany( x1 => e2 , ( x1 , x2 ) => new { x1 , x2 } )
with x3 in e3
…
```
and then translated into (if I understand how transparent identifiers work)
``` c#
from ** in ( ( e1 ) . SelectMany( x1 => e2 , ( x1 , x2 ) => new { x1 , x2 } ) ) . Zip(
e2 , ( * , x3 ) => new { * , x3 })
…
```
And
``` c#
from x1 in e1
with x2 in e2
from x3 in e3
…
```
is first translated into
``` c#
from * in ( e1 ) . Zip( e2 , ( x1 , x2 ) => new { x1 , x2 })
from x3 in e3
…
```
and then translated into
``` c#
from ** in ( ( e1 ) . Zip( e2 , ( x1 , x2 ) => new { x1 , x2 }) ).SelectMany(
* => e3 , ( * , x3 ) => new { * , x3 } )
…
```
| 1.0 | With clause in query expressions (support for zip) - I propose to add a `with` clause to the query expression syntax. This will allow us to zip sequences together using either LINQ syntax.
A query expression with a `with` clause followed by a `select` clause
``` c#
from x1 in e1
with x2 in e2
select v
```
is translated into
``` c#
( e1 ) . Zip( e2, ( x1 , x2 ) => v )
```
A query expression with a `with` clause followed by something other than a `select` clause
``` c#
from x1 in e1
with x2 in e2
…
```
is translated into
``` c#
from * in ( e1 ) . Zip(
e2 , ( x1 , x2 ) => new { x1 , x2 })
…
```
This transformation would happen in section 7.16.2.4 of the spec:
``` c#
from x1 in e1
from x2 in e2
with x3 in e3
…
```
is first translated into
``` c#
from * in ( e1 ) . SelectMany( x1 => e2 , ( x1 , x2 ) => new { x1 , x2 } )
with x3 in e3
…
```
and then translated into (if I understand how transparent identifiers work)
``` c#
from ** in ( ( e1 ) . SelectMany( x1 => e2 , ( x1 , x2 ) => new { x1 , x2 } ) ) . Zip(
e2 , ( * , x3 ) => new { * , x3 })
…
```
And
``` c#
from x1 in e1
with x2 in e2
from x3 in e3
…
```
is first translated into
``` c#
from * in ( e1 ) . Zip( e2 , ( x1 , x2 ) => new { x1 , x2 })
from x3 in e3
…
```
and then translated into
``` c#
from ** in ( ( e1 ) . Zip( e2 , ( x1 , x2 ) => new { x1 , x2 }) ).SelectMany(
* => e3 , ( * , x3 ) => new { * , x3 } )
…
```
| non_main | with clause in query expressions support for zip i propose to add a with clause to the query expression syntax this will allow us to zip sequences together using either linq syntax a query expression with a with clause followed by a select clause c from in with in select v is translated into c zip v a query expression with a with clause followed by something other than a select clause c from in with in … is translated into c from in zip new … this transformation would happen in section of the spec c from in from in with in … is first translated into c from in selectmany new with in … and then translated into if i understand how transparent identifiers work c from in selectmany new zip new … and c from in with in from in … is first translated into c from in zip new from in … and then translated into c from in zip new selectmany new … | 0 |
4,206 | 20,618,209,366 | IssuesEvent | 2022-03-07 15:06:21 | Lissy93/dashy | https://api.github.com/repos/Lissy93/dashy | closed | [BUG] Latest release (2.0.4) is not starting | 🐛 Bug 👤 Awaiting Maintainer Response | <h3 dir="auto">Environment</h3>
<p dir="auto">Self-Hosted (Docker)</p>
<h3 dir="auto">Version</h3>
<p dir="auto">2.0.4</p>
<h3 dir="auto">Describe the problem</h3>
<p dir="auto">With release 2.0.4 container is not becoming ready, its just in "starting" and does not works after hours.</p>
<p dir="auto">I removed my previous configuration (remove the volume) to check if a "fresh restart" works, and after 8 minutes it worked, but never with it.<br>
Trying to check logs there is no output to stdout/stderr, only this:</p>
<blockquote>
<p dir="auto">root@minipc:~/init# docker logs -f dashy<br>
yarn run v1.22.15<br>
$ node server</p>
<p dir="auto">Checking config file against schema...<br>
<g-emoji class="g-emoji" alias="heavy_check_mark" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/2714.png">✔️</g-emoji> Config file is valid, no issues found</p>
<p dir="auto">SSL Not Enabled: Public key not present</p>
<p dir="auto">██████╗ █████╗ ███████╗██╗ ██╗██╗ ██╗<br>
██╔══██╗██╔══██╗██╔════╝██║ ██║╚██╗ ██╔╝<br>
██║ ██║███████║███████╗███████║ ╚████╔╝<br>
██║ ██║██╔══██║╚════██║██╔══██║ ╚██╔╝<br>
██████╔╝██║ ██║███████║██║ ██║ ██║<br>
╚═════╝ ╚═╝ ╚═╝╚══════╝╚═╝ ╚═╝ ╚═╝</p>
<hr>
<p dir="auto">Welcome to Dashy! <g-emoji class="g-emoji" alias="rocket" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/1f680.png">🚀</g-emoji><br>
Your new dashboard is now up and running with Docker</p>
<hr>
<p dir="auto">Using Dashy V-2.0.4. Update Check Complete<br>
<g-emoji class="g-emoji" alias="white_check_mark" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/2705.png">✅</g-emoji> Dashy is Up-to-Date</p>
</blockquote>
<p dir="auto">Couldn't find any path on container with logs...<br>
And couldn't find any new release page, only the releases on github but last update was for 1.9: <a href="https://github.com/Lissy93/dashy/releases">https://github.com/Lissy93/dashy/releases</a></p>
<p dir="auto">I had to go back to 2.0.3 to make it work again.</p>
<h3 dir="auto">Additional info</h3>
<p dir="auto"><em>No response</em></p>
<h3 dir="auto">Please tick the boxes</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You are using a <a href="https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions">supported</a> version of Dashy (check the first two digits of the version number)</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked that this <a href="https://github.com/Lissy93/dashy/issues?q=is%3Aissue">issue hasn't already been raised</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked the <a href="https://github.com/Lissy93/dashy/tree/master/docs#readme">docs</a> and <a href="https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting">troubleshooting</a> guide</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You agree to the <a href="https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct">code of conduct</a></li>
</ul>
<!-- unfurl begin -->
<blockquote> <img src="https://repository-images.githubusercontent.com/343078060/fd44e000-ad15-11eb-8f8a-2feab996e27c" width="48" align="right" /> <div> <strong> <a href="https://github.com/Lissy93/dashy/releases">Releases · Lissy93/dashy</a> </strong> </div> <div>🚀 A self-hosted startpage for your server. Easy to use visual editor, status checking, widgets, themes and tons more! - Lissy93/dashy</div> </blockquote>
<!-- unfurl end --> | True | [BUG] Latest release (2.0.4) is not starting - <h3 dir="auto">Environment</h3>
<p dir="auto">Self-Hosted (Docker)</p>
<h3 dir="auto">Version</h3>
<p dir="auto">2.0.4</p>
<h3 dir="auto">Describe the problem</h3>
<p dir="auto">With release 2.0.4 container is not becoming ready, its just in "starting" and does not works after hours.</p>
<p dir="auto">I removed my previous configuration (remove the volume) to check if a "fresh restart" works, and after 8 minutes it worked, but never with it.<br>
Trying to check logs there is no output to stdout/stderr, only this:</p>
<blockquote>
<p dir="auto">root@minipc:~/init# docker logs -f dashy<br>
yarn run v1.22.15<br>
$ node server</p>
<p dir="auto">Checking config file against schema...<br>
<g-emoji class="g-emoji" alias="heavy_check_mark" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/2714.png">✔️</g-emoji> Config file is valid, no issues found</p>
<p dir="auto">SSL Not Enabled: Public key not present</p>
<p dir="auto">██████╗ █████╗ ███████╗██╗ ██╗██╗ ██╗<br>
██╔══██╗██╔══██╗██╔════╝██║ ██║╚██╗ ██╔╝<br>
██║ ██║███████║███████╗███████║ ╚████╔╝<br>
██║ ██║██╔══██║╚════██║██╔══██║ ╚██╔╝<br>
██████╔╝██║ ██║███████║██║ ██║ ██║<br>
╚═════╝ ╚═╝ ╚═╝╚══════╝╚═╝ ╚═╝ ╚═╝</p>
<hr>
<p dir="auto">Welcome to Dashy! <g-emoji class="g-emoji" alias="rocket" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/1f680.png">🚀</g-emoji><br>
Your new dashboard is now up and running with Docker</p>
<hr>
<p dir="auto">Using Dashy V-2.0.4. Update Check Complete<br>
<g-emoji class="g-emoji" alias="white_check_mark" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/2705.png">✅</g-emoji> Dashy is Up-to-Date</p>
</blockquote>
<p dir="auto">Couldn't find any path on container with logs...<br>
And couldn't find any new release page, only the releases on github but last update was for 1.9: <a href="https://github.com/Lissy93/dashy/releases">https://github.com/Lissy93/dashy/releases</a></p>
<p dir="auto">I had to go back to 2.0.3 to make it work again.</p>
<h3 dir="auto">Additional info</h3>
<p dir="auto"><em>No response</em></p>
<h3 dir="auto">Please tick the boxes</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You are using a <a href="https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions">supported</a> version of Dashy (check the first two digits of the version number)</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked that this <a href="https://github.com/Lissy93/dashy/issues?q=is%3Aissue">issue hasn't already been raised</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked the <a href="https://github.com/Lissy93/dashy/tree/master/docs#readme">docs</a> and <a href="https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting">troubleshooting</a> guide</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You agree to the <a href="https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct">code of conduct</a></li>
</ul>
<!-- unfurl begin -->
<blockquote> <img src="https://repository-images.githubusercontent.com/343078060/fd44e000-ad15-11eb-8f8a-2feab996e27c" width="48" align="right" /> <div> <strong> <a href="https://github.com/Lissy93/dashy/releases">Releases · Lissy93/dashy</a> </strong> </div> <div>🚀 A self-hosted startpage for your server. Easy to use visual editor, status checking, widgets, themes and tons more! - Lissy93/dashy</div> </blockquote>
<!-- unfurl end --> | main | latest release is not starting environment self hosted docker version describe the problem with release container is not becoming ready its just in starting and does not works after hours i removed my previous configuration remove the volume to check if a fresh restart works and after minutes it worked but never with it trying to check logs there is no output to stdout stderr only this root minipc init docker logs f dashy yarn run node server checking config file against schema ssl not enabled public key not present ██████╗ █████╗ ███████╗██╗ ██╗██╗ ██╗ ██╔══██╗██╔══██╗██╔════╝██║ ██║╚██╗ ██╔╝ ██║ ██║███████║███████╗███████║ ╚████╔╝ ██║ ██║██╔══██║╚════██║██╔══██║ ╚██╔╝ ██████╔╝██║ ██║███████║██║ ██║ ██║ ╚═════╝ ╚═╝ ╚═╝╚══════╝╚═╝ ╚═╝ ╚═╝ welcome to dashy g emoji class g emoji alias rocket fallback src your new dashboard is now up and running with docker using dashy v update check complete couldn t find any path on container with logs and couldn t find any new release page only the releases on github but last update was for a href i had to go back to to make it work again additional info no response please tick the boxes you are using a you ve checked that this you ve checked the you agree to the 🚀 a self hosted startpage for your server easy to use visual editor status checking widgets themes and tons more dashy | 1 |
4,031 | 18,836,789,501 | IssuesEvent | 2021-11-11 02:34:29 | prismatic-obloquy/chattor | https://api.github.com/repos/prismatic-obloquy/chattor | closed | Move the Pages site to the docs branch | maintainer-only | Keeping it on its own branch will make it easier to handle, and keep changes to it from bogging down the various pipelines. So... yeah, move it to its own branch and update Pages to point to that. | True | Move the Pages site to the docs branch - Keeping it on its own branch will make it easier to handle, and keep changes to it from bogging down the various pipelines. So... yeah, move it to its own branch and update Pages to point to that. | main | move the pages site to the docs branch keeping it on its own branch will make it easier to handle and keep changes to it from bogging down the various pipelines so yeah move it to its own branch and update pages to point to that | 1 |
2,853 | 10,239,338,254 | IssuesEvent | 2019-08-19 18:00:13 | arcticicestudio/styleguide-javascript | https://api.github.com/repos/arcticicestudio/styleguide-javascript | closed | Monorepo with ESLint packages | context-pkg context-workflow scope-configurability scope-dx scope-maintainability scope-quality scope-stability target-pkg-eslint target-pkg-eslint-base type-epic | ## Current Project State
Currently this repository only contains the actual styleguide documentation while specific projects that implement the guidelines for linters and code style analyzer live in separate repositories. This is the best approach for modularity and a small and clear code base, but it increases the maintenance overhead by 1(n) since changes to the development workflow or toolbox, general project documentations as well as dependency management requires changes in every repository with dedicated tickets/issues and PRs. In particular, Node packages require frequent dependency management due to their fast development cycles to keep up-to-date with the latest package changes like (security) bug fixes.
This styleguide is currently implemented by the [eslint-config-arcticicestudio-base][npm-esl-c-base] and [eslint-config-arcticicestudio][npm-esl-c] Node packages living in their own repositories. The development workflow is clean using most of GitHub's awesome features like project boards, _codeowner_ assignments, issue & PR automation and so on, but changes to one of them often requires actions for the other package too since they are based on each other and they are using the same development tooling and documentation standards.
### Monorepo Comparison
Actually I'm not a supporter when it comes to [monorepos][trbdev-monorepo] and next to the advantages a monorepo also comes with disadvantages:
- **No more scoped code** — The developer experience with Git is clearly worse because commits can contains changes to multiple scopes of the code. Since there are only a “transparent separation” of code, that was previously located in a dedicated repository but is not aggregated into a parent (e.g. `packages`) with other modules, commits can now contain changes to multiple code scopes spread over the entire code base.
- **No more assignment of commits to single modules** — Like described in the bullet point above, commit can contain changes to multiple modules, it is harder to detect which commit targeted a specific module.
- **Steeper learning curve for new contributors** — In a dedicated repository that only hosts a specific module it is easier for new developers to contribute to the project, but in a monorepo they might need to change code in multiple places within other modules or the root code/documentation of the entire project.
- **Uniform version number** — In order to keep conform to [SemVer][], the entire project must use a uniform version number. This means that a module that has not been changed since the last version must also be incremented in order to keep compatible with the other modules.
Using different version numbers prefixed/suffixed with an individual version number **is a not go**, **increases the maintenance overhead** and **and drastically reduces the project overview and quality**! This would result in multiple Git tags on the `master` branch as well as “empty” changelogs and release notes with placeholder logs that only refer to changes of other modules.
## Project Future
Even though there are [disadvantages](#monorepo-disadvantages), a [monorepo][trbdev-monorepo] makes sense **only for specific project modules thar are slightly coupled** and where using dedicated repositories only increases the maintenance overhead **when changes must be reflected in multiple modules anyway**.
In order to reduce the maintenance overhead both Node packages, [eslint-config-arcticicestudio-base][npm-esl-c-base] and [eslint-config-arcticicestudio][npm-esl-c], will migrate into this repository by adapting to [Yarn workspaces][y-d-ws] since they are easily slightly and dependent on each other anyway. This simplifies the development tooling setup and allows to use a unified documentation base as well as a smoother development and testing workflow.
This change also implies that the root of the repository will be the main package for the entire project setup including shared development dependencies, tools and documentations while the packages will only contain specific configurations and (dev)dependencies.
### Scoped Packages
Currently [eslint-config-arcticicestudio-base][npm-esl-c-base] and [eslint-config-arcticicestudio][npm-esl-c] are no [scoped packages][npm-d-scope] but suffixed with `-arcticicestudio*`. To simplify the naming and improving the usage of user/organization specific packages both packages will be scoped to `@arcticicestudio` resulting in the new names `@arcticicestudio/eslint-config-base` and `@arcticicestudio/eslint-config`. They can be used through [ESLint's support for shared configuration with scoped packages][esl-d-shconf_scope].
The currently released public versions [will be deprecated using the `npm deprecate` command][npm-cli-dep] where the provided message will point out to migrate to the new scoped packages.
### Versioning
The style guide itself and all packages will use a shared/fixed/locked version. This helps all packages to keep in sync and ensure the compatibility with the latest style guide version.
### Standard Setup
In order to keep up-to-date with the latest project setup for all _Arctic Ice Studio_ projects, the tools and documentations will be integrated and updated through the following tickets:
- #9 (⊶ 8e992407) „Git ignore and attribute pattern“ — completed ✓
- #10 (⊶ db2a43bc) „Git mail mapping“ — completed ✓
- #11 (⊶ 10253246) „Prettier“ — completed ✓
- #12 (⊶ c21a58a9) „lint-staged“ — completed ✓
- #13 (⊶ b4cac34f) „Husky“ — completed ✓
- #14 (⊶ be122b12) „General repository and package documentations and metadata“ — completed ✓
- #15 (⊶ c25d1efe) „GitHub issue and pull request templates“ — completed ✓
[esl-d-shconf_scope]: https://eslint.org/docs/developer-guide/shareable-configs#npm-scoped-modules
[npm-cli-dep]: https://docs.npmjs.com/cli/deprecate
[npm-d-scope]: https://docs.npmjs.com/about-scopes
[npm-esl-c-base]: https://www.npmjs.com/package/eslint-config-arcticicestudio-base
[npm-esl-c]: https://www.npmjs.com/package/eslint-config-arcticicestudio
[semver]: https://semver.org
[trbdev-monorepo]: https://trunkbaseddevelopment.com/monorepos
[y-d-ws]: https://yarnpkg.com/en/docs/workspaces
| True | Monorepo with ESLint packages - ## Current Project State
Currently this repository only contains the actual styleguide documentation while specific projects that implement the guidelines for linters and code style analyzer live in separate repositories. This is the best approach for modularity and a small and clear code base, but it increases the maintenance overhead by 1(n) since changes to the development workflow or toolbox, general project documentations as well as dependency management requires changes in every repository with dedicated tickets/issues and PRs. In particular, Node packages require frequent dependency management due to their fast development cycles to keep up-to-date with the latest package changes like (security) bug fixes.
This styleguide is currently implemented by the [eslint-config-arcticicestudio-base][npm-esl-c-base] and [eslint-config-arcticicestudio][npm-esl-c] Node packages living in their own repositories. The development workflow is clean using most of GitHub's awesome features like project boards, _codeowner_ assignments, issue & PR automation and so on, but changes to one of them often requires actions for the other package too since they are based on each other and they are using the same development tooling and documentation standards.
### Monorepo Comparison
Actually I'm not a supporter when it comes to [monorepos][trbdev-monorepo] and next to the advantages a monorepo also comes with disadvantages:
- **No more scoped code** — The developer experience with Git is clearly worse because commits can contains changes to multiple scopes of the code. Since there are only a “transparent separation” of code, that was previously located in a dedicated repository but is not aggregated into a parent (e.g. `packages`) with other modules, commits can now contain changes to multiple code scopes spread over the entire code base.
- **No more assignment of commits to single modules** — Like described in the bullet point above, commit can contain changes to multiple modules, it is harder to detect which commit targeted a specific module.
- **Steeper learning curve for new contributors** — In a dedicated repository that only hosts a specific module it is easier for new developers to contribute to the project, but in a monorepo they might need to change code in multiple places within other modules or the root code/documentation of the entire project.
- **Uniform version number** — In order to keep conform to [SemVer][], the entire project must use a uniform version number. This means that a module that has not been changed since the last version must also be incremented in order to keep compatible with the other modules.
Using different version numbers prefixed/suffixed with an individual version number **is a not go**, **increases the maintenance overhead** and **and drastically reduces the project overview and quality**! This would result in multiple Git tags on the `master` branch as well as “empty” changelogs and release notes with placeholder logs that only refer to changes of other modules.
## Project Future
Even though there are [disadvantages](#monorepo-disadvantages), a [monorepo][trbdev-monorepo] makes sense **only for specific project modules thar are slightly coupled** and where using dedicated repositories only increases the maintenance overhead **when changes must be reflected in multiple modules anyway**.
In order to reduce the maintenance overhead both Node packages, [eslint-config-arcticicestudio-base][npm-esl-c-base] and [eslint-config-arcticicestudio][npm-esl-c], will migrate into this repository by adapting to [Yarn workspaces][y-d-ws] since they are easily slightly and dependent on each other anyway. This simplifies the development tooling setup and allows to use a unified documentation base as well as a smoother development and testing workflow.
This change also implies that the root of the repository will be the main package for the entire project setup including shared development dependencies, tools and documentations while the packages will only contain specific configurations and (dev)dependencies.
### Scoped Packages
Currently [eslint-config-arcticicestudio-base][npm-esl-c-base] and [eslint-config-arcticicestudio][npm-esl-c] are no [scoped packages][npm-d-scope] but suffixed with `-arcticicestudio*`. To simplify the naming and improving the usage of user/organization specific packages both packages will be scoped to `@arcticicestudio` resulting in the new names `@arcticicestudio/eslint-config-base` and `@arcticicestudio/eslint-config`. They can be used through [ESLint's support for shared configuration with scoped packages][esl-d-shconf_scope].
The currently released public versions [will be deprecated using the `npm deprecate` command][npm-cli-dep] where the provided message will point out to migrate to the new scoped packages.
### Versioning
The style guide itself and all packages will use a shared/fixed/locked version. This helps all packages to keep in sync and ensure the compatibility with the latest style guide version.
### Standard Setup
In order to keep up-to-date with the latest project setup for all _Arctic Ice Studio_ projects, the tools and documentations will be integrated and updated through the following tickets:
- #9 (⊶ 8e992407) „Git ignore and attribute pattern“ — completed ✓
- #10 (⊶ db2a43bc) „Git mail mapping“ — completed ✓
- #11 (⊶ 10253246) „Prettier“ — completed ✓
- #12 (⊶ c21a58a9) „lint-staged“ — completed ✓
- #13 (⊶ b4cac34f) „Husky“ — completed ✓
- #14 (⊶ be122b12) „General repository and package documentations and metadata“ — completed ✓
- #15 (⊶ c25d1efe) „GitHub issue and pull request templates“ — completed ✓
[esl-d-shconf_scope]: https://eslint.org/docs/developer-guide/shareable-configs#npm-scoped-modules
[npm-cli-dep]: https://docs.npmjs.com/cli/deprecate
[npm-d-scope]: https://docs.npmjs.com/about-scopes
[npm-esl-c-base]: https://www.npmjs.com/package/eslint-config-arcticicestudio-base
[npm-esl-c]: https://www.npmjs.com/package/eslint-config-arcticicestudio
[semver]: https://semver.org
[trbdev-monorepo]: https://trunkbaseddevelopment.com/monorepos
[y-d-ws]: https://yarnpkg.com/en/docs/workspaces
| main | monorepo with eslint packages current project state currently this repository only contains the actual styleguide documentation while specific projects that implement the guidelines for linters and code style analyzer live in separate repositories this is the best approach for modularity and a small and clear code base but it increases the maintenance overhead by n since changes to the development workflow or toolbox general project documentations as well as dependency management requires changes in every repository with dedicated tickets issues and prs in particular node packages require frequent dependency management due to their fast development cycles to keep up to date with the latest package changes like security bug fixes this styleguide is currently implemented by the and node packages living in their own repositories the development workflow is clean using most of github s awesome features like project boards codeowner assignments issue pr automation and so on but changes to one of them often requires actions for the other package too since they are based on each other and they are using the same development tooling and documentation standards monorepo comparison actually i m not a supporter when it comes to and next to the advantages a monorepo also comes with disadvantages no more scoped code — the developer experience with git is clearly worse because commits can contains changes to multiple scopes of the code since there are only a “transparent separation” of code that was previously located in a dedicated repository but is not aggregated into a parent e g packages with other modules commits can now contain changes to multiple code scopes spread over the entire code base no more assignment of commits to single modules — like described in the bullet point above commit can contain changes to multiple modules it is harder to detect which commit targeted a specific module steeper learning curve for new contributors — in a dedicated repository that only hosts a specific module it is easier for new developers to contribute to the project but in a monorepo they might need to change code in multiple places within other modules or the root code documentation of the entire project uniform version number — in order to keep conform to the entire project must use a uniform version number this means that a module that has not been changed since the last version must also be incremented in order to keep compatible with the other modules using different version numbers prefixed suffixed with an individual version number is a not go increases the maintenance overhead and and drastically reduces the project overview and quality this would result in multiple git tags on the master branch as well as “empty” changelogs and release notes with placeholder logs that only refer to changes of other modules project future even though there are monorepo disadvantages a makes sense only for specific project modules thar are slightly coupled and where using dedicated repositories only increases the maintenance overhead when changes must be reflected in multiple modules anyway in order to reduce the maintenance overhead both node packages and will migrate into this repository by adapting to since they are easily slightly and dependent on each other anyway this simplifies the development tooling setup and allows to use a unified documentation base as well as a smoother development and testing workflow this change also implies that the root of the repository will be the main package for the entire project setup including shared development dependencies tools and documentations while the packages will only contain specific configurations and dev dependencies scoped packages currently and are no but suffixed with arcticicestudio to simplify the naming and improving the usage of user organization specific packages both packages will be scoped to arcticicestudio resulting in the new names arcticicestudio eslint config base and arcticicestudio eslint config they can be used through the currently released public versions where the provided message will point out to migrate to the new scoped packages versioning the style guide itself and all packages will use a shared fixed locked version this helps all packages to keep in sync and ensure the compatibility with the latest style guide version standard setup in order to keep up to date with the latest project setup for all arctic ice studio projects the tools and documentations will be integrated and updated through the following tickets ⊶ „git ignore and attribute pattern“ — completed ✓ ⊶ „git mail mapping“ — completed ✓ ⊶ „prettier“ — completed ✓ ⊶ „lint staged“ — completed ✓ ⊶ „husky“ — completed ✓ ⊶ „general repository and package documentations and metadata“ — completed ✓ ⊶ „github issue and pull request templates“ — completed ✓ | 1 |
1,031 | 4,827,518,233 | IssuesEvent | 2016-11-07 13:53:11 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | Consul module idempotency | affects_2.3 feature_idea waiting_on_maintainer | Hey guys! I wonder if anyone can provide a bit of insight into this comment:
https://github.com/ansible/ansible-modules-extras/blob/fdb4a58f975fc4df4b22157b97428b0b274a0d3d/clustering/consul.py#L33-L38
Was hoping I could help further that effort along on either the consul or the ansible side, but after some searching im not able to find much news about this upcoming api :(
Maybe @sgargan can help? Thanks so much in advance!
| True | Consul module idempotency - Hey guys! I wonder if anyone can provide a bit of insight into this comment:
https://github.com/ansible/ansible-modules-extras/blob/fdb4a58f975fc4df4b22157b97428b0b274a0d3d/clustering/consul.py#L33-L38
Was hoping I could help further that effort along on either the consul or the ansible side, but after some searching im not able to find much news about this upcoming api :(
Maybe @sgargan can help? Thanks so much in advance!
| main | consul module idempotency hey guys i wonder if anyone can provide a bit of insight into this comment was hoping i could help further that effort along on either the consul or the ansible side but after some searching im not able to find much news about this upcoming api maybe sgargan can help thanks so much in advance | 1 |
3,387 | 13,157,647,443 | IssuesEvent | 2020-08-10 13:06:45 | RapidField/solid-instruments | https://api.github.com/repos/RapidField/solid-instruments | closed | Add registration aggregate for Entity Framework dependencies. | Category-Feature Source-Maintainer Stage-3-InProgress Subcategory-Functionality Subsystem-DataAccess Tag-AddReleaseNote Verdict-Pending Version-1.0.26 WindowForDelivery-2021-Q1 | # Feature Request
This issue represents a suggestion for new **Solid Instruments** functionality.
## Overview
Extend the **Autofac** and native .NET dependency injection abstractions to expose single-line registration options for **Entity Framework** support dependencies.
## Statement of work
The following list describes the work to be done and defines acceptance criteria for the feature.
1. Create a new project named `RapidField.SolidInstruments.DataAccess.Autofac.Ef`.
2. Create a new project named `RapidField.SolidInstruments.DataAccess.DotNetNative.Ef`.
3. Create an aggregate dependency registration method for **Entity Framework** support via **Autofac**.
4. Create an aggregate dependency registration method for **Entity Framework** support via native .NET IoC.
5. Extend the documentation to cover the new components.
## Additional information
## Revision control plan
**Solid Instruments** uses the [**RapidField Revision Control Workflow**](https://github.com/RapidField/solid-instruments/blob/master/CONTRIBUTING.md#revision-control-strategy). Individual contributors should follow the branching plan below when working on this issue.
- `master` is the pull request target for
- `release/v1.0.26-preview1`, which is the pull request target for
- `develop`, which is the pull request target for
- `feature/000304-ef-registration`, which is the pull request target for contributing user branches, which should be named using the pattern
- `user/{username}/00304-ef-registration` | True | Add registration aggregate for Entity Framework dependencies. - # Feature Request
This issue represents a suggestion for new **Solid Instruments** functionality.
## Overview
Extend the **Autofac** and native .NET dependency injection abstractions to expose single-line registration options for **Entity Framework** support dependencies.
## Statement of work
The following list describes the work to be done and defines acceptance criteria for the feature.
1. Create a new project named `RapidField.SolidInstruments.DataAccess.Autofac.Ef`.
2. Create a new project named `RapidField.SolidInstruments.DataAccess.DotNetNative.Ef`.
3. Create an aggregate dependency registration method for **Entity Framework** support via **Autofac**.
4. Create an aggregate dependency registration method for **Entity Framework** support via native .NET IoC.
5. Extend the documentation to cover the new components.
## Additional information
## Revision control plan
**Solid Instruments** uses the [**RapidField Revision Control Workflow**](https://github.com/RapidField/solid-instruments/blob/master/CONTRIBUTING.md#revision-control-strategy). Individual contributors should follow the branching plan below when working on this issue.
- `master` is the pull request target for
- `release/v1.0.26-preview1`, which is the pull request target for
- `develop`, which is the pull request target for
- `feature/000304-ef-registration`, which is the pull request target for contributing user branches, which should be named using the pattern
- `user/{username}/00304-ef-registration` | main | add registration aggregate for entity framework dependencies feature request this issue represents a suggestion for new solid instruments functionality overview extend the autofac and native net dependency injection abstractions to expose single line registration options for entity framework support dependencies statement of work the following list describes the work to be done and defines acceptance criteria for the feature create a new project named rapidfield solidinstruments dataaccess autofac ef create a new project named rapidfield solidinstruments dataaccess dotnetnative ef create an aggregate dependency registration method for entity framework support via autofac create an aggregate dependency registration method for entity framework support via native net ioc extend the documentation to cover the new components additional information revision control plan solid instruments uses the individual contributors should follow the branching plan below when working on this issue master is the pull request target for release which is the pull request target for develop which is the pull request target for feature ef registration which is the pull request target for contributing user branches which should be named using the pattern user username ef registration | 1 |
3,610 | 14,598,191,656 | IssuesEvent | 2020-12-20 23:41:00 | spack/spack | https://api.github.com/repos/spack/spack | closed | Maintainers bots and actions | actions feature maintainers | I've been trying to think of ways to increase our number of packages with official maintainers. Here's what I came up with.
1. Add a GitHub action to automatically add package `maintainers` as reviewers of PRs. This was attempted in #12269, but later removed because we never got it working. I wanted to open a dedicated issue on this so we remember to try again.
2. Add a GitHub action to automatically add package `maintainers` as assignees for issues. This is a little harder, as we can't directly see which package is modified, it needs to be inferred from the issue title. I propose we standardize on `package: ` or `[package]` in the issue title.
3. Add a bot with a standardized "Would you like to add yourself as a maintainer? Here's how. Here's why." message that comments on all PRs for packages without a maintainer. This could get a little annoying, but hopefully it will convince people to add themselves as maintainers. | True | Maintainers bots and actions - I've been trying to think of ways to increase our number of packages with official maintainers. Here's what I came up with.
1. Add a GitHub action to automatically add package `maintainers` as reviewers of PRs. This was attempted in #12269, but later removed because we never got it working. I wanted to open a dedicated issue on this so we remember to try again.
2. Add a GitHub action to automatically add package `maintainers` as assignees for issues. This is a little harder, as we can't directly see which package is modified, it needs to be inferred from the issue title. I propose we standardize on `package: ` or `[package]` in the issue title.
3. Add a bot with a standardized "Would you like to add yourself as a maintainer? Here's how. Here's why." message that comments on all PRs for packages without a maintainer. This could get a little annoying, but hopefully it will convince people to add themselves as maintainers. | main | maintainers bots and actions i ve been trying to think of ways to increase our number of packages with official maintainers here s what i came up with add a github action to automatically add package maintainers as reviewers of prs this was attempted in but later removed because we never got it working i wanted to open a dedicated issue on this so we remember to try again add a github action to automatically add package maintainers as assignees for issues this is a little harder as we can t directly see which package is modified it needs to be inferred from the issue title i propose we standardize on package or in the issue title add a bot with a standardized would you like to add yourself as a maintainer here s how here s why message that comments on all prs for packages without a maintainer this could get a little annoying but hopefully it will convince people to add themselves as maintainers | 1 |
91,390 | 10,719,823,294 | IssuesEvent | 2019-10-26 13:15:57 | locustio/locust | https://api.github.com/repos/locustio/locust | closed | Possible typo in docs | documentation | <!-- For general questions about how to use Locust, use either the Slack link provided in the Readme or [ask a question on Stack Overflow](https://stackoverflow.com/questions/ask) tagged Locust.-->
### Description of issue / feature request
In this section of the docs: https://docs.locust.io/en/stable/writing-a-locustfile.html#manually-controlling-if-a-request-should-be-considered-successful-or-a-failure `client` is referenced *not* as an attribute. Everywhere else it's `self.client`. Is this a typo?
### Expected behavior
Code examples should reference `self.client`
### Actual behavior
Some code examples reference just `client`
### Environment settings (for bug reports)
Current edition of docs (0.9.0)
### Steps to reproduce (for bug reports)
N/A | 1.0 | Possible typo in docs - <!-- For general questions about how to use Locust, use either the Slack link provided in the Readme or [ask a question on Stack Overflow](https://stackoverflow.com/questions/ask) tagged Locust.-->
### Description of issue / feature request
In this section of the docs: https://docs.locust.io/en/stable/writing-a-locustfile.html#manually-controlling-if-a-request-should-be-considered-successful-or-a-failure `client` is referenced *not* as an attribute. Everywhere else it's `self.client`. Is this a typo?
### Expected behavior
Code examples should reference `self.client`
### Actual behavior
Some code examples reference just `client`
### Environment settings (for bug reports)
Current edition of docs (0.9.0)
### Steps to reproduce (for bug reports)
N/A | non_main | possible typo in docs description of issue feature request in this section of the docs client is referenced not as an attribute everywhere else it s self client is this a typo expected behavior code examples should reference self client actual behavior some code examples reference just client environment settings for bug reports current edition of docs steps to reproduce for bug reports n a | 0 |
5,883 | 32,024,765,347 | IssuesEvent | 2023-09-22 08:03:22 | camunda/zeebe | https://api.github.com/repos/camunda/zeebe | closed | `expression-language` module shouldn't depend on `scheduler` | kind/toil good first issue area/maintainability component/scheduler | **Description**
The `zeebe-expression-language` depends on the `scheduler` module because it accepts an `ActorClock` as it's source of time. This is a bit inelegant and we should make changes so that `zeebe-expression-language` doesn't have a direct dependency on `scheduler`, for example by injecting `ZeebeFeelEngineClock` instead of `ActorClock`.
See related discussion here: https://github.com/camunda/zeebe/pull/9597#discussion_r907021048
| True | `expression-language` module shouldn't depend on `scheduler` - **Description**
The `zeebe-expression-language` depends on the `scheduler` module because it accepts an `ActorClock` as it's source of time. This is a bit inelegant and we should make changes so that `zeebe-expression-language` doesn't have a direct dependency on `scheduler`, for example by injecting `ZeebeFeelEngineClock` instead of `ActorClock`.
See related discussion here: https://github.com/camunda/zeebe/pull/9597#discussion_r907021048
| main | expression language module shouldn t depend on scheduler description the zeebe expression language depends on the scheduler module because it accepts an actorclock as it s source of time this is a bit inelegant and we should make changes so that zeebe expression language doesn t have a direct dependency on scheduler for example by injecting zeebefeelengineclock instead of actorclock see related discussion here | 1 |
756,710 | 26,482,661,893 | IssuesEvent | 2023-01-17 15:43:24 | LiteLDev/LiteLoaderBDS | https://api.github.com/repos/LiteLDev/LiteLoaderBDS | closed | `JsonConfigFile`行为异常 | type: bug module: script engine priority: high | ### 异常模块
ScriptEngine (脚本引擎)
### 操作系统
Windows Server 2022
### LiteLoader 版本
2.9.1 Actions#123 779364e
### BDS 版本
1.19.51.01
### 发生了什么?
`new JsonConfigFile`创建的是文件夹
### 复现此问题的步骤
_No response_
### 有关的日志/输出
_No response_
### 插件列表
_No response_ | 1.0 | `JsonConfigFile`行为异常 - ### 异常模块
ScriptEngine (脚本引擎)
### 操作系统
Windows Server 2022
### LiteLoader 版本
2.9.1 Actions#123 779364e
### BDS 版本
1.19.51.01
### 发生了什么?
`new JsonConfigFile`创建的是文件夹
### 复现此问题的步骤
_No response_
### 有关的日志/输出
_No response_
### 插件列表
_No response_ | non_main | jsonconfigfile 行为异常 异常模块 scriptengine 脚本引擎 操作系统 windows server liteloader 版本 actions bds 版本 发生了什么 new jsonconfigfile 创建的是文件夹 复现此问题的步骤 no response 有关的日志 输出 no response 插件列表 no response | 0 |
613,774 | 19,098,071,189 | IssuesEvent | 2021-11-29 18:56:21 | kubeapps/kubeapps | https://api.github.com/repos/kubeapps/kubeapps | opened | Installed app not found right after installing a Carvel package | kind/bug component/dashboard priority/high size/S | Parent issue: https://github.com/kubeapps/kubeapps/issues/3849
When a package is installed, we got redirected to the InstalledPackage view, which is great. The problem is that Kapp hasn't created the `App` resource yet. Alternatives (I'd go with 2)
1. Auto-retry from the UI in N attempts
- follow-up idea: should we add exp. backoff retries to every get-alike request to avoid transient API connectivity issues?
2. Check the `App` resource exists in the backend before sending the OK response back. | 1.0 | Installed app not found right after installing a Carvel package - Parent issue: https://github.com/kubeapps/kubeapps/issues/3849
When a package is installed, we got redirected to the InstalledPackage view, which is great. The problem is that Kapp hasn't created the `App` resource yet. Alternatives (I'd go with 2)
1. Auto-retry from the UI in N attempts
- follow-up idea: should we add exp. backoff retries to every get-alike request to avoid transient API connectivity issues?
2. Check the `App` resource exists in the backend before sending the OK response back. | non_main | installed app not found right after installing a carvel package parent issue when a package is installed we got redirected to the installedpackage view which is great the problem is that kapp hasn t created the app resource yet alternatives i d go with auto retry from the ui in n attempts follow up idea should we add exp backoff retries to every get alike request to avoid transient api connectivity issues check the app resource exists in the backend before sending the ok response back | 0 |
3,159 | 12,220,380,904 | IssuesEvent | 2020-05-02 01:17:14 | short-d/short | https://api.github.com/repos/short-d/short | opened | [Refactor] Generate random alias if empty or malicious string is passed for alias from frontend | maintainability | **What is frustrating you?**
When you specify an empty string custom alias (currently impossible from the frontend but a custom form can be crafted to do this) and click "Create Short Link", it is passed as an empty string into the backend. This will end up producing a short link of empty string alias, which is undesired.
**Your solution**
The backend should interpret an empty string alias as a request to randomly generate an alias. Additionally, if malicious input is detected for the alias, also generate a random alias and log that malicious input was attempt (but do NOT provide any feedback to user that malicious input was attempted, just provide the same output as if they provided an empty string)
**Additional context**

In the DB, if you look up the record it'll look like this

An alternative solution was considered to check for empty string and pass `null` as the argument instead to backend, but malicious user can easily remove this check from the frontend code.

| True | [Refactor] Generate random alias if empty or malicious string is passed for alias from frontend - **What is frustrating you?**
When you specify an empty string custom alias (currently impossible from the frontend but a custom form can be crafted to do this) and click "Create Short Link", it is passed as an empty string into the backend. This will end up producing a short link of empty string alias, which is undesired.
**Your solution**
The backend should interpret an empty string alias as a request to randomly generate an alias. Additionally, if malicious input is detected for the alias, also generate a random alias and log that malicious input was attempt (but do NOT provide any feedback to user that malicious input was attempted, just provide the same output as if they provided an empty string)
**Additional context**

In the DB, if you look up the record it'll look like this

An alternative solution was considered to check for empty string and pass `null` as the argument instead to backend, but malicious user can easily remove this check from the frontend code.

| main | generate random alias if empty or malicious string is passed for alias from frontend what is frustrating you when you specify an empty string custom alias currently impossible from the frontend but a custom form can be crafted to do this and click create short link it is passed as an empty string into the backend this will end up producing a short link of empty string alias which is undesired your solution the backend should interpret an empty string alias as a request to randomly generate an alias additionally if malicious input is detected for the alias also generate a random alias and log that malicious input was attempt but do not provide any feedback to user that malicious input was attempted just provide the same output as if they provided an empty string additional context in the db if you look up the record it ll look like this an alternative solution was considered to check for empty string and pass null as the argument instead to backend but malicious user can easily remove this check from the frontend code | 1 |
186,955 | 14,426,868,445 | IssuesEvent | 2020-12-06 00:28:37 | kalexmills/github-vet-tests-dec2020 | https://api.github.com/repos/kalexmills/github-vet-tests-dec2020 | closed | munnerz/k8s-api-pager-demo: vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go; 6 LoC | fresh test tiny vendored |
Found a possible issue in [munnerz/k8s-api-pager-demo](https://www.github.com/munnerz/k8s-api-pager-demo) at [vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go](https://github.com/munnerz/k8s-api-pager-demo/blob/c4470dbbc304d9f165df0b77ae15da170b07b16a/vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go#L66-L71)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to v at line 68 may start a goroutine
[Click here to see the code in its original context.](https://github.com/munnerz/k8s-api-pager-demo/blob/c4470dbbc304d9f165df0b77ae15da170b07b16a/vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go#L66-L71)
<details>
<summary>Click here to show the 6 line(s) of Go which triggered the analyzer.</summary>
```go
for k, v := range cases {
actual := ErrorToAPIStatus(k)
if !reflect.DeepEqual(actual, &v) {
t.Errorf("%s: Expected %#v, Got %#v", k, v, actual)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: c4470dbbc304d9f165df0b77ae15da170b07b16a
| 1.0 | munnerz/k8s-api-pager-demo: vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go; 6 LoC -
Found a possible issue in [munnerz/k8s-api-pager-demo](https://www.github.com/munnerz/k8s-api-pager-demo) at [vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go](https://github.com/munnerz/k8s-api-pager-demo/blob/c4470dbbc304d9f165df0b77ae15da170b07b16a/vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go#L66-L71)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to v at line 68 may start a goroutine
[Click here to see the code in its original context.](https://github.com/munnerz/k8s-api-pager-demo/blob/c4470dbbc304d9f165df0b77ae15da170b07b16a/vendor/k8s.io/apiserver/pkg/endpoints/handlers/responsewriters/status_test.go#L66-L71)
<details>
<summary>Click here to show the 6 line(s) of Go which triggered the analyzer.</summary>
```go
for k, v := range cases {
actual := ErrorToAPIStatus(k)
if !reflect.DeepEqual(actual, &v) {
t.Errorf("%s: Expected %#v, Got %#v", k, v, actual)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: c4470dbbc304d9f165df0b77ae15da170b07b16a
| non_main | munnerz api pager demo vendor io apiserver pkg endpoints handlers responsewriters status test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to v at line may start a goroutine click here to show the line s of go which triggered the analyzer go for k v range cases actual errortoapistatus k if reflect deepequal actual v t errorf s expected v got v k v actual leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
4,895 | 25,138,767,445 | IssuesEvent | 2022-11-09 21:01:50 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Keyboard fails to move active cell after clicking a column header or row header | type: bug work: frontend status: ready restricted: maintainers | ## Steps to reproduce
1. Go to a table page.
1. Click on a cell to select it.
1. Press `Up`, `Down`, `Left`, `Right`, `Tab` and `Shift+Tab`.
1. Observe the active cell to move in response to the keyboard. Good.
1. Click on a column header to select all the cells within the column.
1. Press those keyboard keys again.
1. Expect the active cell to move as before.
1. Instead, observe that the active cell does not move in response to the keyboard.
| True | Keyboard fails to move active cell after clicking a column header or row header - ## Steps to reproduce
1. Go to a table page.
1. Click on a cell to select it.
1. Press `Up`, `Down`, `Left`, `Right`, `Tab` and `Shift+Tab`.
1. Observe the active cell to move in response to the keyboard. Good.
1. Click on a column header to select all the cells within the column.
1. Press those keyboard keys again.
1. Expect the active cell to move as before.
1. Instead, observe that the active cell does not move in response to the keyboard.
| main | keyboard fails to move active cell after clicking a column header or row header steps to reproduce go to a table page click on a cell to select it press up down left right tab and shift tab observe the active cell to move in response to the keyboard good click on a column header to select all the cells within the column press those keyboard keys again expect the active cell to move as before instead observe that the active cell does not move in response to the keyboard | 1 |
88,953 | 10,584,732,093 | IssuesEvent | 2019-10-08 15:59:44 | Astrotomic/stancy | https://api.github.com/repos/Astrotomic/stancy | opened | Add nginx config for static serving | documentation | Add a default nginx config to serve static HTML files and Laravel app as fallback. | 1.0 | Add nginx config for static serving - Add a default nginx config to serve static HTML files and Laravel app as fallback. | non_main | add nginx config for static serving add a default nginx config to serve static html files and laravel app as fallback | 0 |
5,215 | 26,464,347,001 | IssuesEvent | 2023-01-16 21:18:51 | bazelbuild/intellij | https://api.github.com/repos/bazelbuild/intellij | closed | Flag --incompatible_disable_starlark_host_transitions will break Android Studio Plugin Google in Bazel 7.0 | type: bug product: Android Studio topic: bazel awaiting-maintainer | Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking Android Studio Plugin Google. Please migrate to fix this and unblock the flip of this flag.
The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032).
Please check the following CI builds for build and test results:
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fce-4eab-8e4d-97ca84a465e6)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fd1-4cd4-b59c-0f6fa4745159)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fcb-46b6-83a5-aade011f8ca4)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fd7-4793-b4ce-1611e0db64a8)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fd4-41f9-ab4a-4591aa0ba68b)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fda-4163-8bea-6bdc9864800e)
Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything.
If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration. | True | Flag --incompatible_disable_starlark_host_transitions will break Android Studio Plugin Google in Bazel 7.0 - Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking Android Studio Plugin Google. Please migrate to fix this and unblock the flip of this flag.
The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032).
Please check the following CI builds for build and test results:
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fce-4eab-8e4d-97ca84a465e6)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fd1-4cd4-b59c-0f6fa4745159)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fcb-46b6-83a5-aade011f8ca4)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fd7-4793-b4ce-1611e0db64a8)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fd4-41f9-ab4a-4591aa0ba68b)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4fda-4163-8bea-6bdc9864800e)
Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything.
If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration. | main | flag incompatible disable starlark host transitions will break android studio plugin google in bazel incompatible flag incompatible disable starlark host transitions will be enabled by default in the next major release bazel thus breaking android studio plugin google please migrate to fix this and unblock the flip of this flag the flag is documented here please check the following ci builds for build and test results never heard of incompatible flags before we have that explains everything if you have any questions please file an issue in | 1 |
256,982 | 22,139,545,815 | IssuesEvent | 2022-06-03 04:41:20 | brave/brave-ios | https://api.github.com/repos/brave/brave-ios | closed | Manual test run for `1.39` on iPhone running `iOS 15` | QA/Yes release-notes/exclude QA Pass - iPhone tests iPhone |
## Installer
- [x] Check that installer is close to the size of the last release
- [x] Check the Brave version in About and make sure it is EXACTLY as expected
## Data
- [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, history, etc.)
- [x] Verify that cookies from the previous build are preserved after upgrade
- [x] Verify saved passwords are retained after upgrade
- [x] Verify stats are retained after upgrade
- [x] Verify sync chain created in the previous version is still retained on upgrade
- [x] Verify per-site settings are preserved after upgrade
## Bookmarks
- [x] Verify that creating a bookmark works
- [x] Verify that tapping a bookmark from bookmark manager loads the bookmark
- [x] Verify that deleting a bookmark works
- [x] Verify that creating a bookmark folder works
- [x] Verify that creating a bookmark inside the created folder works
- [x] Verify that you can add a bookmark directly inside a bookmark folder
- [x] Verify that you can delete a bookmark in edit mode
- [x] Verify that you can delete a bookmark folder with bookmarks inside
- [x] Verify adding a bookmark domain, subpaths is retained and you are successfully able to visit the domain subpath in a new tab
## Favourites
- [x] Verify editing favourites and changing their URLs updates the favicons accordingly
- [x] Verify that you can remove favourites
- [x] Verify that you can add new favourites from the `Share with...` menu
## Context menus
- [x] Make sure context menu items in the URL bar work
- [x] Make sure context menu items on content work with no selected text
- [x] Make sure context menu items on content work with selected text
- [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable)
- [x] Verify you can open links in a new tab or private tab. Ensure you can open links in background tabs.
## Downloads
- [x] Verify that you can save an image from a site
- [x] Verify that you are able to save a GIF image
## Reader Mode
- [x] Visit `http://theverge.com`, open any article, verify the reader mode icon is shown in the URL bar
- [x] Verify tapping on the reader mode icon opens the article in reader mode
- [x] Edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article
## Gestures
- [x] Verify pinch to zoom in/out gestures work on `https://www.homedepot.com`
- [x] Verify that navigating to a different origin resets the zoom
- [x] Swipe back and forward to navigate, verify this works as expected
## Password Managers
- [x] Verify tapping on 1Password on the slide-out keyboard launches 1Password app and you're able to select the stored credentials
- [x] Verify tapping on Bitwarden password manager in the autofill field launches the app and auto-fills the stored data
## Shields Settings
- [x] Enable all switches in `Settings` and visit a site and disable `Block Scripts`. Kill and relaunch the app and verify if the site Shields settings are retained
## Browser Lock
- [x] Verify browser lock enables device passcode to lock/unlock the browser
- [x] Verify swipe up/swipe down with browser in focus doesn't ask for device passcode
- [x] Verify cancel passcode shows unlock message
- [x] Remove the app from memory and relaunch, enter a wrong passcode, the browser should not be unlocked
- [x] Verify cancel fingerprint confirmation/face unlock shows enter passcode window when fingerprint/face unlock is set up on the device
## Brave Rewards/Ads
- [x] Verify wallet is auto-created after enabling Rewards
- [x] Verify when you click on the Brave Rewards panel while on a site, the panel displays if the site is verified or not
- [x] Verify ads are only shown while the app is being used
- [x] Verify tapping on an ad notification shows the landing page
- [x] Verify inline-content ads show on Brave News and are correctly redeemed from the server
- [x] Verify `view`,`clicked`, `landed`, and `dismiss` states are logged based on the action
## Sync
- [x] Verify you are able to join Sync chain by scanning the QR code
- [x] Verify you are able to join Sync chain using code words
- [x] Verify you are able to create a Sync chain on the device and add other devices to the chain via QR code/code words
- [x] Verify that bookmarks from other devices on the chain show up on the mobile device after sync completes
- [x] Verify newly created bookmarks get synced to all devices on the Sync chain
- [x] Verify existing bookmarks before joining Sync chain also gets sync'd to all devices on the Sync chain
- [x] Verify history sync is disabled by default. Enabling it starts sync'ing history items as well (limited to 200 entries).
- [x] Verify sync works on an upgraded profile and new bookmarks added post-upgrade sync across devices on the chain
- [x] Verify you can create a standalone Sync chain with one device
## Playlist & Autoplay
- [x] Verify Playlist is enabled by default on a clean install
- [x] Verify when a site with audio/video is detected, `Playlist` button is shown in URL bar and a notification badge on the menu
- [x] Verify clicking the `Playlist` button adds audio/video to playlist and initiates offline play
- [x] Verify able to add audio/video to playlist via share menu
- [x] Verify able to add audio/video to playlist via long press context menu
- [x] Verify Autoplay is disabled by default
- [x] Verify with Autoplay enabled, able to play audio/video automatically
- [x] Verify with background audio enabled, able to play audio/video with device locked or app in background
## Brave News
- [x] Verify Brave News opt-in is shown on a clean install
- [x] Verify once opted-in Brave News starts showing articles
- [ ] Verify promoted content is shown on Brave News
- [x] Verify inline-content ads are shown on Brave News
- [x] Verify able to add custom RSS feed to the list of sources
- [x] Verify able to disable content from a particular publisher by long-press on the card
## Bravery settings
- [ ] Check that HTTPS Everywhere works by loading `http://https-everywhere.badssl.com/`
- [ ] Turning HTTPS Everywhere off or shields off both disable the redirect to `https://https-everywhere.badssl.com/`
- [x] Check that block ad and unblock ad works on `http://slashdot.org`
- [x] Check that toggling to blocking and allow ads works as expected
- [x] Verify that tapping through a cert error in `https://badssl.com` works
- [ ] Verify that Safe Browsing works `https://www.raisegame.com`
- [x] Enable `Block Scripts` globally from `Settings`. Visit `https://twitter.com`, nothing should load. Tap on `Shields` and disable `Block Scripts`; the page should load properly.
- [x] Verify that default Brave Shields settings take effect on pages with no site settings
### Fingerprint Tests
- [x] Verify that turning on fingerprinting protection in preferences shows `3 fingerprints blocked` at `https://browserleaks.com/canvas`. Verify that turning it off in the Bravery menu shows `0 fingerprints blocked`.
- [x] Verify that the audio fingerprint is blocked at `https://audiofingerprint.openwpm.com` when fingerprinting protection is on
- [x] Verify that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address when `Block all fingerprinting protection` is on
## Content tests
- [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again.
- [x] Open an email on `https://mail.google.com` and tap on a link. Make sure it works.
- [x] Verify that PDF is loaded over `HTTPS` at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf`
- [x] Verify that PDF is loaded over `HTTP` at `http://www.pdf995.com/samples/pdf.pdf`
- [x] Verify that `https://mixed-script.badssl.com` shows up as grey, not red (no mixed-content scripts are run)
- [x] Verify that search results from `https://startpage.com` open in a new tab (due to target being _blank_)
## WebAuthn
- [x] Verify browser prompts for security key when trying to log in to accounts
- [x] Verify you are able to successfully authenticate an account using security key when prompted
- [x] Verify quickly connecting and disconnecting doesn't cause issues authenticating using security keys
- [x] Verify you are able to use security keys on both normal and private tabs
- [ ] Verify you are able to use security keys to authenticate using NFC
## App linker
- [x] Long-press on a link in the Twitter app to get the share picker, choose Brave. Verify Brave doesn't crash after opening the link.
## Session storage
- [x] Verify that tabs restore when closed, including active tab
| 1.0 | Manual test run for `1.39` on iPhone running `iOS 15` -
## Installer
- [x] Check that installer is close to the size of the last release
- [x] Check the Brave version in About and make sure it is EXACTLY as expected
## Data
- [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, history, etc.)
- [x] Verify that cookies from the previous build are preserved after upgrade
- [x] Verify saved passwords are retained after upgrade
- [x] Verify stats are retained after upgrade
- [x] Verify sync chain created in the previous version is still retained on upgrade
- [x] Verify per-site settings are preserved after upgrade
## Bookmarks
- [x] Verify that creating a bookmark works
- [x] Verify that tapping a bookmark from bookmark manager loads the bookmark
- [x] Verify that deleting a bookmark works
- [x] Verify that creating a bookmark folder works
- [x] Verify that creating a bookmark inside the created folder works
- [x] Verify that you can add a bookmark directly inside a bookmark folder
- [x] Verify that you can delete a bookmark in edit mode
- [x] Verify that you can delete a bookmark folder with bookmarks inside
- [x] Verify adding a bookmark domain, subpaths is retained and you are successfully able to visit the domain subpath in a new tab
## Favourites
- [x] Verify editing favourites and changing their URLs updates the favicons accordingly
- [x] Verify that you can remove favourites
- [x] Verify that you can add new favourites from the `Share with...` menu
## Context menus
- [x] Make sure context menu items in the URL bar work
- [x] Make sure context menu items on content work with no selected text
- [x] Make sure context menu items on content work with selected text
- [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable)
- [x] Verify you can open links in a new tab or private tab. Ensure you can open links in background tabs.
## Downloads
- [x] Verify that you can save an image from a site
- [x] Verify that you are able to save a GIF image
## Reader Mode
- [x] Visit `http://theverge.com`, open any article, verify the reader mode icon is shown in the URL bar
- [x] Verify tapping on the reader mode icon opens the article in reader mode
- [x] Edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article
## Gestures
- [x] Verify pinch to zoom in/out gestures work on `https://www.homedepot.com`
- [x] Verify that navigating to a different origin resets the zoom
- [x] Swipe back and forward to navigate, verify this works as expected
## Password Managers
- [x] Verify tapping on 1Password on the slide-out keyboard launches 1Password app and you're able to select the stored credentials
- [x] Verify tapping on Bitwarden password manager in the autofill field launches the app and auto-fills the stored data
## Shields Settings
- [x] Enable all switches in `Settings` and visit a site and disable `Block Scripts`. Kill and relaunch the app and verify if the site Shields settings are retained
## Browser Lock
- [x] Verify browser lock enables device passcode to lock/unlock the browser
- [x] Verify swipe up/swipe down with browser in focus doesn't ask for device passcode
- [x] Verify cancel passcode shows unlock message
- [x] Remove the app from memory and relaunch, enter a wrong passcode, the browser should not be unlocked
- [x] Verify cancel fingerprint confirmation/face unlock shows enter passcode window when fingerprint/face unlock is set up on the device
## Brave Rewards/Ads
- [x] Verify wallet is auto-created after enabling Rewards
- [x] Verify when you click on the Brave Rewards panel while on a site, the panel displays if the site is verified or not
- [x] Verify ads are only shown while the app is being used
- [x] Verify tapping on an ad notification shows the landing page
- [x] Verify inline-content ads show on Brave News and are correctly redeemed from the server
- [x] Verify `view`,`clicked`, `landed`, and `dismiss` states are logged based on the action
## Sync
- [x] Verify you are able to join Sync chain by scanning the QR code
- [x] Verify you are able to join Sync chain using code words
- [x] Verify you are able to create a Sync chain on the device and add other devices to the chain via QR code/code words
- [x] Verify that bookmarks from other devices on the chain show up on the mobile device after sync completes
- [x] Verify newly created bookmarks get synced to all devices on the Sync chain
- [x] Verify existing bookmarks before joining Sync chain also gets sync'd to all devices on the Sync chain
- [x] Verify history sync is disabled by default. Enabling it starts sync'ing history items as well (limited to 200 entries).
- [x] Verify sync works on an upgraded profile and new bookmarks added post-upgrade sync across devices on the chain
- [x] Verify you can create a standalone Sync chain with one device
## Playlist & Autoplay
- [x] Verify Playlist is enabled by default on a clean install
- [x] Verify when a site with audio/video is detected, `Playlist` button is shown in URL bar and a notification badge on the menu
- [x] Verify clicking the `Playlist` button adds audio/video to playlist and initiates offline play
- [x] Verify able to add audio/video to playlist via share menu
- [x] Verify able to add audio/video to playlist via long press context menu
- [x] Verify Autoplay is disabled by default
- [x] Verify with Autoplay enabled, able to play audio/video automatically
- [x] Verify with background audio enabled, able to play audio/video with device locked or app in background
## Brave News
- [x] Verify Brave News opt-in is shown on a clean install
- [x] Verify once opted-in Brave News starts showing articles
- [ ] Verify promoted content is shown on Brave News
- [x] Verify inline-content ads are shown on Brave News
- [x] Verify able to add custom RSS feed to the list of sources
- [x] Verify able to disable content from a particular publisher by long-press on the card
## Bravery settings
- [ ] Check that HTTPS Everywhere works by loading `http://https-everywhere.badssl.com/`
- [ ] Turning HTTPS Everywhere off or shields off both disable the redirect to `https://https-everywhere.badssl.com/`
- [x] Check that block ad and unblock ad works on `http://slashdot.org`
- [x] Check that toggling to blocking and allow ads works as expected
- [x] Verify that tapping through a cert error in `https://badssl.com` works
- [ ] Verify that Safe Browsing works `https://www.raisegame.com`
- [x] Enable `Block Scripts` globally from `Settings`. Visit `https://twitter.com`, nothing should load. Tap on `Shields` and disable `Block Scripts`; the page should load properly.
- [x] Verify that default Brave Shields settings take effect on pages with no site settings
### Fingerprint Tests
- [x] Verify that turning on fingerprinting protection in preferences shows `3 fingerprints blocked` at `https://browserleaks.com/canvas`. Verify that turning it off in the Bravery menu shows `0 fingerprints blocked`.
- [x] Verify that the audio fingerprint is blocked at `https://audiofingerprint.openwpm.com` when fingerprinting protection is on
- [x] Verify that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address when `Block all fingerprinting protection` is on
## Content tests
- [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again.
- [x] Open an email on `https://mail.google.com` and tap on a link. Make sure it works.
- [x] Verify that PDF is loaded over `HTTPS` at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf`
- [x] Verify that PDF is loaded over `HTTP` at `http://www.pdf995.com/samples/pdf.pdf`
- [x] Verify that `https://mixed-script.badssl.com` shows up as grey, not red (no mixed-content scripts are run)
- [x] Verify that search results from `https://startpage.com` open in a new tab (due to target being _blank_)
## WebAuthn
- [x] Verify browser prompts for security key when trying to log in to accounts
- [x] Verify you are able to successfully authenticate an account using security key when prompted
- [x] Verify quickly connecting and disconnecting doesn't cause issues authenticating using security keys
- [x] Verify you are able to use security keys on both normal and private tabs
- [ ] Verify you are able to use security keys to authenticate using NFC
## App linker
- [x] Long-press on a link in the Twitter app to get the share picker, choose Brave. Verify Brave doesn't crash after opening the link.
## Session storage
- [x] Verify that tabs restore when closed, including active tab
| non_main | manual test run for on iphone running ios installer check that installer is close to the size of the last release check the brave version in about and make sure it is exactly as expected data verify that data from the previous build appears in the updated build as expected bookmarks history etc verify that cookies from the previous build are preserved after upgrade verify saved passwords are retained after upgrade verify stats are retained after upgrade verify sync chain created in the previous version is still retained on upgrade verify per site settings are preserved after upgrade bookmarks verify that creating a bookmark works verify that tapping a bookmark from bookmark manager loads the bookmark verify that deleting a bookmark works verify that creating a bookmark folder works verify that creating a bookmark inside the created folder works verify that you can add a bookmark directly inside a bookmark folder verify that you can delete a bookmark in edit mode verify that you can delete a bookmark folder with bookmarks inside verify adding a bookmark domain subpaths is retained and you are successfully able to visit the domain subpath in a new tab favourites verify editing favourites and changing their urls updates the favicons accordingly verify that you can remove favourites verify that you can add new favourites from the share with menu context menus make sure context menu items in the url bar work make sure context menu items on content work with no selected text make sure context menu items on content work with selected text make sure context menu items on content work inside an editable control input textarea or contenteditable verify you can open links in a new tab or private tab ensure you can open links in background tabs downloads verify that you can save an image from a site verify that you are able to save a gif image reader mode visit open any article verify the reader mode icon is shown in the url bar verify tapping on the reader mode icon opens the article in reader mode edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article gestures verify pinch to zoom in out gestures work on verify that navigating to a different origin resets the zoom swipe back and forward to navigate verify this works as expected password managers verify tapping on on the slide out keyboard launches app and you re able to select the stored credentials verify tapping on bitwarden password manager in the autofill field launches the app and auto fills the stored data shields settings enable all switches in settings and visit a site and disable block scripts kill and relaunch the app and verify if the site shields settings are retained browser lock verify browser lock enables device passcode to lock unlock the browser verify swipe up swipe down with browser in focus doesn t ask for device passcode verify cancel passcode shows unlock message remove the app from memory and relaunch enter a wrong passcode the browser should not be unlocked verify cancel fingerprint confirmation face unlock shows enter passcode window when fingerprint face unlock is set up on the device brave rewards ads verify wallet is auto created after enabling rewards verify when you click on the brave rewards panel while on a site the panel displays if the site is verified or not verify ads are only shown while the app is being used verify tapping on an ad notification shows the landing page verify inline content ads show on brave news and are correctly redeemed from the server verify view clicked landed and dismiss states are logged based on the action sync verify you are able to join sync chain by scanning the qr code verify you are able to join sync chain using code words verify you are able to create a sync chain on the device and add other devices to the chain via qr code code words verify that bookmarks from other devices on the chain show up on the mobile device after sync completes verify newly created bookmarks get synced to all devices on the sync chain verify existing bookmarks before joining sync chain also gets sync d to all devices on the sync chain verify history sync is disabled by default enabling it starts sync ing history items as well limited to entries verify sync works on an upgraded profile and new bookmarks added post upgrade sync across devices on the chain verify you can create a standalone sync chain with one device playlist autoplay verify playlist is enabled by default on a clean install verify when a site with audio video is detected playlist button is shown in url bar and a notification badge on the menu verify clicking the playlist button adds audio video to playlist and initiates offline play verify able to add audio video to playlist via share menu verify able to add audio video to playlist via long press context menu verify autoplay is disabled by default verify with autoplay enabled able to play audio video automatically verify with background audio enabled able to play audio video with device locked or app in background brave news verify brave news opt in is shown on a clean install verify once opted in brave news starts showing articles verify promoted content is shown on brave news verify inline content ads are shown on brave news verify able to add custom rss feed to the list of sources verify able to disable content from a particular publisher by long press on the card bravery settings check that https everywhere works by loading turning https everywhere off or shields off both disable the redirect to check that block ad and unblock ad works on check that toggling to blocking and allow ads works as expected verify that tapping through a cert error in works verify that safe browsing works enable block scripts globally from settings visit nothing should load tap on shields and disable block scripts the page should load properly verify that default brave shields settings take effect on pages with no site settings fingerprint tests verify that turning on fingerprinting protection in preferences shows fingerprints blocked at verify that turning it off in the bravery menu shows fingerprints blocked verify that the audio fingerprint is blocked at when fingerprinting protection is on verify that doesn t leak ip address when block all fingerprinting protection is on content tests go to and make sure that the password can be saved make sure the saved password is auto populated when you visit the site again open an email on and tap on a link make sure it works verify that pdf is loaded over https at verify that pdf is loaded over http at verify that shows up as grey not red no mixed content scripts are run verify that search results from open in a new tab due to target being blank webauthn verify browser prompts for security key when trying to log in to accounts verify you are able to successfully authenticate an account using security key when prompted verify quickly connecting and disconnecting doesn t cause issues authenticating using security keys verify you are able to use security keys on both normal and private tabs verify you are able to use security keys to authenticate using nfc app linker long press on a link in the twitter app to get the share picker choose brave verify brave doesn t crash after opening the link session storage verify that tabs restore when closed including active tab | 0 |
5,640 | 28,369,224,291 | IssuesEvent | 2023-04-12 15:45:54 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Allow pasting from the system clipboard into Mathesar cells | type: enhancement work: frontend status: ready restricted: maintainers | ## Current behavior
- Mathesar allows cell data to be _copied_ to the system clipboard (implemented in #2773)
- There is no _paste_ functionality though.
## Desired behavior
- "Paste" is quite a bit more complex than "copy". The behavior is specified in detail within our [Clipboard Interaction specs](https://wiki.mathesar.org/en/design/specs/clipboard-interactions)
- Note that pasting requires reading clipboard data associated with our Mathesar-specific MIME type, but we are not yet _writing_ that representation to the clipboard when copying. So, to fully implement pasting, we'll need to also update the copy action to write that new representation too.
- The spec leaves some room for improvisation during implementation:
- The deserialization of TSV data has lots of edge cases and I think it's worth attempting to delegate this logic to a 3rd party lib, if we can find a suitable one. The copy operation doesn't yet fully implement our specified serialization logic. We have #2811 to track that, and I think it will probably make sense to tackle that issue concurrently with this one.
| True | Allow pasting from the system clipboard into Mathesar cells - ## Current behavior
- Mathesar allows cell data to be _copied_ to the system clipboard (implemented in #2773)
- There is no _paste_ functionality though.
## Desired behavior
- "Paste" is quite a bit more complex than "copy". The behavior is specified in detail within our [Clipboard Interaction specs](https://wiki.mathesar.org/en/design/specs/clipboard-interactions)
- Note that pasting requires reading clipboard data associated with our Mathesar-specific MIME type, but we are not yet _writing_ that representation to the clipboard when copying. So, to fully implement pasting, we'll need to also update the copy action to write that new representation too.
- The spec leaves some room for improvisation during implementation:
- The deserialization of TSV data has lots of edge cases and I think it's worth attempting to delegate this logic to a 3rd party lib, if we can find a suitable one. The copy operation doesn't yet fully implement our specified serialization logic. We have #2811 to track that, and I think it will probably make sense to tackle that issue concurrently with this one.
| main | allow pasting from the system clipboard into mathesar cells current behavior mathesar allows cell data to be copied to the system clipboard implemented in there is no paste functionality though desired behavior paste is quite a bit more complex than copy the behavior is specified in detail within our note that pasting requires reading clipboard data associated with our mathesar specific mime type but we are not yet writing that representation to the clipboard when copying so to fully implement pasting we ll need to also update the copy action to write that new representation too the spec leaves some room for improvisation during implementation the deserialization of tsv data has lots of edge cases and i think it s worth attempting to delegate this logic to a party lib if we can find a suitable one the copy operation doesn t yet fully implement our specified serialization logic we have to track that and i think it will probably make sense to tackle that issue concurrently with this one | 1 |
231,870 | 17,759,771,178 | IssuesEvent | 2021-08-29 13:23:57 | matteodelabre/vnsee | https://api.github.com/repos/matteodelabre/vnsee | closed | Build fails looking for <filesystem> header | documentation | I've probably just missed some basic configuration option, but 0d94c7f39b817c3cb0ea921fc41e9ee702f90295 adds the `#include <filesystem>` header, which causes my build to fail:
```
[ 39%] Building CXX object CMakeFiles/vnsee.dir/src/rmioc/device.cpp.o
/home/steven/Documents/projects/remarkable/vnsee/src/rmioc/device.cpp:6:10: fatal error: filesystem: No such file or directory
#include <filesystem>
^~~~~~~~~~~~
compilation terminated.
```
Do I need to enable C++17 support somewhere?
I've got the recommended toolchain installed: `oecore-x86_64-cortexa9hf-neon-toolchain-zero-gravitas-1.8-23.9.2019`
```
$ arm-oe-linux-gnueabi-gcc --version
arm-oe-linux-gnueabi-gcc (GCC) 7.3.0
Copyright (C) 2017 Free Software Foundation, Inc.
``` | 1.0 | Build fails looking for <filesystem> header - I've probably just missed some basic configuration option, but 0d94c7f39b817c3cb0ea921fc41e9ee702f90295 adds the `#include <filesystem>` header, which causes my build to fail:
```
[ 39%] Building CXX object CMakeFiles/vnsee.dir/src/rmioc/device.cpp.o
/home/steven/Documents/projects/remarkable/vnsee/src/rmioc/device.cpp:6:10: fatal error: filesystem: No such file or directory
#include <filesystem>
^~~~~~~~~~~~
compilation terminated.
```
Do I need to enable C++17 support somewhere?
I've got the recommended toolchain installed: `oecore-x86_64-cortexa9hf-neon-toolchain-zero-gravitas-1.8-23.9.2019`
```
$ arm-oe-linux-gnueabi-gcc --version
arm-oe-linux-gnueabi-gcc (GCC) 7.3.0
Copyright (C) 2017 Free Software Foundation, Inc.
``` | non_main | build fails looking for header i ve probably just missed some basic configuration option but adds the include header which causes my build to fail building cxx object cmakefiles vnsee dir src rmioc device cpp o home steven documents projects remarkable vnsee src rmioc device cpp fatal error filesystem no such file or directory include compilation terminated do i need to enable c support somewhere i ve got the recommended toolchain installed oecore neon toolchain zero gravitas arm oe linux gnueabi gcc version arm oe linux gnueabi gcc gcc copyright c free software foundation inc | 0 |
1,975 | 6,694,173,755 | IssuesEvent | 2017-10-10 00:05:35 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | closed | Places: show working time range | Improvement Internal Maintainer Approved | Could we also show opening/closing time on queries like "McDonalds nearby".
---
IA Page: http://duck.co/ia/view/maps_places
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @nilnilnil
| True | Places: show working time range - Could we also show opening/closing time on queries like "McDonalds nearby".
---
IA Page: http://duck.co/ia/view/maps_places
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @nilnilnil
| main | places show working time range could we also show opening closing time on queries like mcdonalds nearby ia page nilnilnil | 1 |
3,265 | 12,424,994,645 | IssuesEvent | 2020-05-24 14:22:37 | dappkit/aviondb | https://api.github.com/repos/dappkit/aviondb | closed | Typescript Interfaces & ESLint Config | maintainance | It'd be great to see the AvionDB environment fully integrated with Typescript (interfaces for each class etc, strict types for method arguments etc).
I also noticed a lot of variables use `var` instead of `const` or `let`. An ESLint config could help enforce some requirements (such as only use `const` or `let` instead of `var` , only allow arrow functions etc) | True | Typescript Interfaces & ESLint Config - It'd be great to see the AvionDB environment fully integrated with Typescript (interfaces for each class etc, strict types for method arguments etc).
I also noticed a lot of variables use `var` instead of `const` or `let`. An ESLint config could help enforce some requirements (such as only use `const` or `let` instead of `var` , only allow arrow functions etc) | main | typescript interfaces eslint config it d be great to see the aviondb environment fully integrated with typescript interfaces for each class etc strict types for method arguments etc i also noticed a lot of variables use var instead of const or let an eslint config could help enforce some requirements such as only use const or let instead of var only allow arrow functions etc | 1 |
5,206 | 26,457,601,785 | IssuesEvent | 2023-01-16 15:18:56 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Add support for self-referential links within Link Table modal | type: enhancement work: frontend status: ready restricted: maintainers | Description and discussion pending
| True | Add support for self-referential links within Link Table modal - Description and discussion pending
| main | add support for self referential links within link table modal description and discussion pending | 1 |
1,699 | 6,574,376,520 | IssuesEvent | 2017-09-11 12:39:45 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | ec2_asg does not work when using replace_all_instances option | affects_2.2 aws bug_report cloud waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_asg
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
OSX, local_action
##### SUMMARY
When using the option replace_all_instances within below task:
```
- name: Configure Auto Scaling Groups
local_action:
module: ec2_asg
name: "{{ item.name }}"
desired_capacity: "{{ item.desired_capacity }}"
launch_config_name: "{{ item.launch_config }}"
replace_all_instances: yes
region: sa-east-1
become: false
with_items:
- { name: "{{ ondemand_asg_name }}", desired_capacity: "{{ ondemand_desired_capacity }}", launch_config: "{{ ondemand_lc }}" }
#- { name: "{{ spot_asg_name }}", desired_capacity: "{{ spot_desired_capacity }}", launch_config: "{{ spot_lc }}" }
register: asg_output
when: configasg is defined
```
I can see using aws console that the group size is indeed changing and that the launch_configuration was replaced as expected, but while the module is waiting for the newly launched instances to be InService it fails with the following message and leaves the asg with unwanted minimum, maximum and desired values as well as the effect of using replace_all_instances is not reached since there is old instances left running with older launch_configuration.
```
Using module file /Library/Python/2.7/site-packages/ansible/modules/core/cloud/amazon/ec2_asg.py
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478796730.65-629465686899 `" && echo ansible-tmp-1478796730.65-629465686899="` echo $HOME/.ansible/tmp/ansible-tmp-1478796730.65-629465686899 `" ) && sleep 0'
<localhost> PUT /var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/tmpS2gR2m TO /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ec2_asg.py
<localhost> EXEC /bin/sh -c 'chmod u+x /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ec2_asg.py && sleep 0'
<localhost> EXEC /bin/sh -c '/usr/bin/python /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ec2_asg.py; rm -rf "/Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 875, in <module>
main()
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 869, in main
replace_changed, asg_properties=replace(connection, module)
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 644, in replace
break_early, desired_size, term_instances = terminate_batch(connection, module, i, instances, False)
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 730, in terminate_batch
as_group.update()
File "/Library/Python/2.7/site-packages/boto/ec2/autoscale/group.py", line 282, in update
return self.connection._update_group('UpdateAutoScalingGroup', self)
File "/Library/Python/2.7/site-packages/boto/ec2/autoscale/__init__.py", line 183, in _update_group
return self.get_object(op, params, Request)
File "/Library/Python/2.7/site-packages/boto/connection.py", line 1208, in get_object
raise self.ResponseError(response.status, response.reason, body)
boto.exception.BotoServerError: BotoServerError: 400 Bad Request
<ErrorResponse xmlns="http://autoscaling.amazonaws.com/doc/2011-01-01/">
<Error>
<Type>Sender</Type>
<Code>MalformedInput</Code>
</Error>
<RequestId>3968c89c-a766-11e6-8907-3df0a5528d6b</RequestId>
</ErrorResponse>
failed: [127.0.0.1 -> localhost] (item={u'desired_capacity': u'2', u'launch_config': u'zupme-gateway-autoscaling-2.25.0-spot-2016-11-10-10_32', u'name': u'zupme-gateway-autoscaling-spot'}) => {
"failed": true,
"invocation": {
"module_name": "ec2_asg"
},
"item": {
"desired_capacity": "2",
"launch_config": "zupme-gateway-autoscaling-2.25.0-spot-2016-11-10-10_32",
"name": "zupme-gateway-autoscaling-spot"
},
"module_stderr": "Traceback (most recent call last):\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 875, in <module>\n main()\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 869, in main\n replace_changed, asg_properties=replace(connection, module)\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 644, in replace\n break_early, desired_size, term_instances = terminate_batch(connection, module, i, instances, False)\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 730, in terminate_batch\n as_group.update()\n File \"/Library/Python/2.7/site-packages/boto/ec2/autoscale/group.py\", line 282, in update\n return self.connection._update_group('UpdateAutoScalingGroup', self)\n File \"/Library/Python/2.7/site-packages/boto/ec2/autoscale/__init__.py\", line 183, in _update_group\n return self.get_object(op, params, Request)\n File \"/Library/Python/2.7/site-packages/boto/connection.py\", line 1208, in get_object\n raise self.ResponseError(response.status, response.reason, body)\nboto.exception.BotoServerError: BotoServerError: 400 Bad Request\n<ErrorResponse xmlns=\"http://autoscaling.amazonaws.com/doc/2011-01-01/\">\n <Error>\n <Type>Sender</Type>\n <Code>MalformedInput</Code>\n </Error>\n <RequestId>3968c89c-a766-11e6-8907-3df0a5528d6b</RequestId>\n</ErrorResponse>\n\n",
"module_stdout": "",
"msg": "MODULE FAILURE"
}
```
##### STEPS TO REPRODUCE
Create a ec2_asg task with replace_all_instances and try to run it.
##### EXPECTED RESULTS
launch_configuration changed and instances with older launch_configuration should be terminated.
##### ACTUAL RESULTS
ec2_asg leaves the asg with unwanted minimum, maximum and desired values as well as the effect of using replace_all_instances is not reached since there are old instances left running with older launch_configuration.
| True | ec2_asg does not work when using replace_all_instances option - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_asg
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
OSX, local_action
##### SUMMARY
When using the option replace_all_instances within below task:
```
- name: Configure Auto Scaling Groups
local_action:
module: ec2_asg
name: "{{ item.name }}"
desired_capacity: "{{ item.desired_capacity }}"
launch_config_name: "{{ item.launch_config }}"
replace_all_instances: yes
region: sa-east-1
become: false
with_items:
- { name: "{{ ondemand_asg_name }}", desired_capacity: "{{ ondemand_desired_capacity }}", launch_config: "{{ ondemand_lc }}" }
#- { name: "{{ spot_asg_name }}", desired_capacity: "{{ spot_desired_capacity }}", launch_config: "{{ spot_lc }}" }
register: asg_output
when: configasg is defined
```
I can see using aws console that the group size is indeed changing and that the launch_configuration was replaced as expected, but while the module is waiting for the newly launched instances to be InService it fails with the following message and leaves the asg with unwanted minimum, maximum and desired values as well as the effect of using replace_all_instances is not reached since there is old instances left running with older launch_configuration.
```
Using module file /Library/Python/2.7/site-packages/ansible/modules/core/cloud/amazon/ec2_asg.py
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478796730.65-629465686899 `" && echo ansible-tmp-1478796730.65-629465686899="` echo $HOME/.ansible/tmp/ansible-tmp-1478796730.65-629465686899 `" ) && sleep 0'
<localhost> PUT /var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/tmpS2gR2m TO /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ec2_asg.py
<localhost> EXEC /bin/sh -c 'chmod u+x /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ec2_asg.py && sleep 0'
<localhost> EXEC /bin/sh -c '/usr/bin/python /Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/ec2_asg.py; rm -rf "/Users/underguiz/.ansible/tmp/ansible-tmp-1478796730.65-629465686899/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 875, in <module>
main()
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 869, in main
replace_changed, asg_properties=replace(connection, module)
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 644, in replace
break_early, desired_size, term_instances = terminate_batch(connection, module, i, instances, False)
File "/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py", line 730, in terminate_batch
as_group.update()
File "/Library/Python/2.7/site-packages/boto/ec2/autoscale/group.py", line 282, in update
return self.connection._update_group('UpdateAutoScalingGroup', self)
File "/Library/Python/2.7/site-packages/boto/ec2/autoscale/__init__.py", line 183, in _update_group
return self.get_object(op, params, Request)
File "/Library/Python/2.7/site-packages/boto/connection.py", line 1208, in get_object
raise self.ResponseError(response.status, response.reason, body)
boto.exception.BotoServerError: BotoServerError: 400 Bad Request
<ErrorResponse xmlns="http://autoscaling.amazonaws.com/doc/2011-01-01/">
<Error>
<Type>Sender</Type>
<Code>MalformedInput</Code>
</Error>
<RequestId>3968c89c-a766-11e6-8907-3df0a5528d6b</RequestId>
</ErrorResponse>
failed: [127.0.0.1 -> localhost] (item={u'desired_capacity': u'2', u'launch_config': u'zupme-gateway-autoscaling-2.25.0-spot-2016-11-10-10_32', u'name': u'zupme-gateway-autoscaling-spot'}) => {
"failed": true,
"invocation": {
"module_name": "ec2_asg"
},
"item": {
"desired_capacity": "2",
"launch_config": "zupme-gateway-autoscaling-2.25.0-spot-2016-11-10-10_32",
"name": "zupme-gateway-autoscaling-spot"
},
"module_stderr": "Traceback (most recent call last):\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 875, in <module>\n main()\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 869, in main\n replace_changed, asg_properties=replace(connection, module)\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 644, in replace\n break_early, desired_size, term_instances = terminate_batch(connection, module, i, instances, False)\n File \"/var/folders/_n/vjpj21ld7nzdxl50d46w6bjm0000gn/T/ansible_2eKuRg/ansible_module_ec2_asg.py\", line 730, in terminate_batch\n as_group.update()\n File \"/Library/Python/2.7/site-packages/boto/ec2/autoscale/group.py\", line 282, in update\n return self.connection._update_group('UpdateAutoScalingGroup', self)\n File \"/Library/Python/2.7/site-packages/boto/ec2/autoscale/__init__.py\", line 183, in _update_group\n return self.get_object(op, params, Request)\n File \"/Library/Python/2.7/site-packages/boto/connection.py\", line 1208, in get_object\n raise self.ResponseError(response.status, response.reason, body)\nboto.exception.BotoServerError: BotoServerError: 400 Bad Request\n<ErrorResponse xmlns=\"http://autoscaling.amazonaws.com/doc/2011-01-01/\">\n <Error>\n <Type>Sender</Type>\n <Code>MalformedInput</Code>\n </Error>\n <RequestId>3968c89c-a766-11e6-8907-3df0a5528d6b</RequestId>\n</ErrorResponse>\n\n",
"module_stdout": "",
"msg": "MODULE FAILURE"
}
```
##### STEPS TO REPRODUCE
Create a ec2_asg task with replace_all_instances and try to run it.
##### EXPECTED RESULTS
launch_configuration changed and instances with older launch_configuration should be terminated.
##### ACTUAL RESULTS
ec2_asg leaves the asg with unwanted minimum, maximum and desired values as well as the effect of using replace_all_instances is not reached since there are old instances left running with older launch_configuration.
| main | asg does not work when using replace all instances option issue type bug report component name asg ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides os environment osx local action summary when using the option replace all instances within below task name configure auto scaling groups local action module asg name item name desired capacity item desired capacity launch config name item launch config replace all instances yes region sa east become false with items name ondemand asg name desired capacity ondemand desired capacity launch config ondemand lc name spot asg name desired capacity spot desired capacity launch config spot lc register asg output when configasg is defined i can see using aws console that the group size is indeed changing and that the launch configuration was replaced as expected but while the module is waiting for the newly launched instances to be inservice it fails with the following message and leaves the asg with unwanted minimum maximum and desired values as well as the effect of using replace all instances is not reached since there is old instances left running with older launch configuration using module file library python site packages ansible modules core cloud amazon asg py exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders n t to users underguiz ansible tmp ansible tmp asg py exec bin sh c chmod u x users underguiz ansible tmp ansible tmp users underguiz ansible tmp ansible tmp asg py sleep exec bin sh c usr bin python users underguiz ansible tmp ansible tmp asg py rm rf users underguiz ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file var folders n t ansible ansible module asg py line in main file var folders n t ansible ansible module asg py line in main replace changed asg properties replace connection module file var folders n t ansible ansible module asg py line in replace break early desired size term instances terminate batch connection module i instances false file var folders n t ansible ansible module asg py line in terminate batch as group update file library python site packages boto autoscale group py line in update return self connection update group updateautoscalinggroup self file library python site packages boto autoscale init py line in update group return self get object op params request file library python site packages boto connection py line in get object raise self responseerror response status response reason body boto exception botoservererror botoservererror bad request errorresponse xmlns sender malformedinput failed item u desired capacity u u launch config u zupme gateway autoscaling spot u name u zupme gateway autoscaling spot failed true invocation module name asg item desired capacity launch config zupme gateway autoscaling spot name zupme gateway autoscaling spot module stderr traceback most recent call last n file var folders n t ansible ansible module asg py line in n main n file var folders n t ansible ansible module asg py line in main n replace changed asg properties replace connection module n file var folders n t ansible ansible module asg py line in replace n break early desired size term instances terminate batch connection module i instances false n file var folders n t ansible ansible module asg py line in terminate batch n as group update n file library python site packages boto autoscale group py line in update n return self connection update group updateautoscalinggroup self n file library python site packages boto autoscale init py line in update group n return self get object op params request n file library python site packages boto connection py line in get object n raise self responseerror response status response reason body nboto exception botoservererror botoservererror bad request n n sender n malformedinput n n n n n module stdout msg module failure steps to reproduce create a asg task with replace all instances and try to run it expected results launch configuration changed and instances with older launch configuration should be terminated actual results asg leaves the asg with unwanted minimum maximum and desired values as well as the effect of using replace all instances is not reached since there are old instances left running with older launch configuration | 1 |
3,603 | 14,546,271,435 | IssuesEvent | 2020-12-15 20:58:12 | microsoft/DirectXTK | https://api.github.com/repos/microsoft/DirectXTK | closed | Add vcpkg installation instructions | maintainence | `directxtk `is available as a port in [vcpkg](https://github.com/microsoft/vcpkg), a C++ library manager that simplifies installation for `directxtk `and other project dependencies. Documenting the install process here will help users get started by providing a single set of commands to build `directxtk `, ready to be included in their projects.
We also test whether our library ports build in various configurations (dynamic, static) on various platforms (Windows: x86, x64, arm, uwp) to keep a wide coverage for users.
Note: Currently it only supports dynamic build on Windows.
I'm a maintainer for vcpkg, and [here is what the port script looks like](https://github.com/microsoft/vcpkg/blob/master/ports/directxtk/portfile.cmake). We try to keep the library maintained as close as possible to the original library.
I'm trying to add vcpkg installation instructions but I couldn't find the related docs in code item except for on `Wiki `page.
Could you please help add the following contents to this port?
Installing and building directxtk via vcpkg
You can download and install directxtk using the [vcpkg](https://github.com/Microsoft/vcpkg) dependency manager:
git clone https://github.com/Microsoft/vcpkg.git
cd vcpkg
./bootstrap-vcpkg.sh
./vcpkg integrate install
vcpkg install openmvs
The directxtk port in vcpkg is kept up to date by Microsoft team members and community contributors. If the version is out of date, please [create an issue or pull request](https://github.com/Microsoft/vcpkg) on the vcpkg repository.
Thanks,
Nancy | True | Add vcpkg installation instructions - `directxtk `is available as a port in [vcpkg](https://github.com/microsoft/vcpkg), a C++ library manager that simplifies installation for `directxtk `and other project dependencies. Documenting the install process here will help users get started by providing a single set of commands to build `directxtk `, ready to be included in their projects.
We also test whether our library ports build in various configurations (dynamic, static) on various platforms (Windows: x86, x64, arm, uwp) to keep a wide coverage for users.
Note: Currently it only supports dynamic build on Windows.
I'm a maintainer for vcpkg, and [here is what the port script looks like](https://github.com/microsoft/vcpkg/blob/master/ports/directxtk/portfile.cmake). We try to keep the library maintained as close as possible to the original library.
I'm trying to add vcpkg installation instructions but I couldn't find the related docs in code item except for on `Wiki `page.
Could you please help add the following contents to this port?
Installing and building directxtk via vcpkg
You can download and install directxtk using the [vcpkg](https://github.com/Microsoft/vcpkg) dependency manager:
git clone https://github.com/Microsoft/vcpkg.git
cd vcpkg
./bootstrap-vcpkg.sh
./vcpkg integrate install
vcpkg install openmvs
The directxtk port in vcpkg is kept up to date by Microsoft team members and community contributors. If the version is out of date, please [create an issue or pull request](https://github.com/Microsoft/vcpkg) on the vcpkg repository.
Thanks,
Nancy | main | add vcpkg installation instructions directxtk is available as a port in a c library manager that simplifies installation for directxtk and other project dependencies documenting the install process here will help users get started by providing a single set of commands to build directxtk ready to be included in their projects we also test whether our library ports build in various configurations dynamic static on various platforms windows arm uwp to keep a wide coverage for users note currently it only supports dynamic build on windows i m a maintainer for vcpkg and we try to keep the library maintained as close as possible to the original library i m trying to add vcpkg installation instructions but i couldn t find the related docs in code item except for on wiki page could you please help add the following contents to this port installing and building directxtk via vcpkg you can download and install directxtk using the dependency manager git clone cd vcpkg bootstrap vcpkg sh vcpkg integrate install vcpkg install openmvs the directxtk port in vcpkg is kept up to date by microsoft team members and community contributors if the version is out of date please on the vcpkg repository thanks nancy | 1 |
3,797 | 16,238,951,623 | IssuesEvent | 2021-05-07 06:53:18 | walbourn/directx-vs-templates | https://api.github.com/repos/walbourn/directx-vs-templates | closed | Remove support for Windows 10 SDK (14393) or earlier | maintainence | There are a few places in the templates where I use this guard.
```
#if defined(NTDDI_WIN10_RS2) && (NTDDI_VERSION >= NTDDI_WIN10_RS2)
```
This is specifically to support the Windows 10 SDK (14393) which was the last one supported for VS 2015. Since I no longer support VS 2015, I can remove these guards to simplify the code. | True | Remove support for Windows 10 SDK (14393) or earlier - There are a few places in the templates where I use this guard.
```
#if defined(NTDDI_WIN10_RS2) && (NTDDI_VERSION >= NTDDI_WIN10_RS2)
```
This is specifically to support the Windows 10 SDK (14393) which was the last one supported for VS 2015. Since I no longer support VS 2015, I can remove these guards to simplify the code. | main | remove support for windows sdk or earlier there are a few places in the templates where i use this guard if defined ntddi ntddi version ntddi this is specifically to support the windows sdk which was the last one supported for vs since i no longer support vs i can remove these guards to simplify the code | 1 |
736,617 | 25,480,242,070 | IssuesEvent | 2022-11-25 19:35:14 | bcgov/entity | https://api.github.com/repos/bcgov/entity | closed | PPR Search by Name - Only first letter of middle name is matched for "exact match" purposes | bug Priority3 Assets | ## Description
When I perform a PPR search by name, only the first letter of the middle name is matched for the purpose of identifying "exact matches" on the search results screen. This is a function of the search algorithm, but the reason why is not known at this time (it may have been a holdover from how the legacy application handled middle names).
## Options
- [ ] Option 1: Update the search algorithm to show exact matches for full middle name
- [ ] Option 2: Update UI to give user feedback alerting user that only first letter of middle name is used for exact matches
## Notes
- Ops ticket: https://github.com/bcgov-registries/ops-support/issues/1261
- Related ticket from previous bug fix: #8557 | 1.0 | PPR Search by Name - Only first letter of middle name is matched for "exact match" purposes - ## Description
When I perform a PPR search by name, only the first letter of the middle name is matched for the purpose of identifying "exact matches" on the search results screen. This is a function of the search algorithm, but the reason why is not known at this time (it may have been a holdover from how the legacy application handled middle names).
## Options
- [ ] Option 1: Update the search algorithm to show exact matches for full middle name
- [ ] Option 2: Update UI to give user feedback alerting user that only first letter of middle name is used for exact matches
## Notes
- Ops ticket: https://github.com/bcgov-registries/ops-support/issues/1261
- Related ticket from previous bug fix: #8557 | non_main | ppr search by name only first letter of middle name is matched for exact match purposes description when i perform a ppr search by name only the first letter of the middle name is matched for the purpose of identifying exact matches on the search results screen this is a function of the search algorithm but the reason why is not known at this time it may have been a holdover from how the legacy application handled middle names options option update the search algorithm to show exact matches for full middle name option update ui to give user feedback alerting user that only first letter of middle name is used for exact matches notes ops ticket related ticket from previous bug fix | 0 |
3,740 | 15,707,387,800 | IssuesEvent | 2021-03-26 18:48:23 | precice/precice | https://api.github.com/repos/precice/precice | closed | Extend CI to macOS | maintainability | We got the hardware.
Next steps:
* Add a macOS build to the workflow
* Setup and use a local runner | True | Extend CI to macOS - We got the hardware.
Next steps:
* Add a macOS build to the workflow
* Setup and use a local runner | main | extend ci to macos we got the hardware next steps add a macos build to the workflow setup and use a local runner | 1 |
1,486 | 2,860,124,295 | IssuesEvent | 2015-06-03 14:29:49 | Graylog2/graylog2-web-interface | https://api.github.com/repos/Graylog2/graylog2-web-interface | opened | Trying to access a deleted saved search should redirect you nicely | usability | On 1.1.0-rc.3, trying to access a saved search that was deleted (or doesn't exist) results in a error page. Instead of this, we should properly redirect to a normal search and explain that the search was not found. | True | Trying to access a deleted saved search should redirect you nicely - On 1.1.0-rc.3, trying to access a saved search that was deleted (or doesn't exist) results in a error page. Instead of this, we should properly redirect to a normal search and explain that the search was not found. | non_main | trying to access a deleted saved search should redirect you nicely on rc trying to access a saved search that was deleted or doesn t exist results in a error page instead of this we should properly redirect to a normal search and explain that the search was not found | 0 |
212,468 | 16,452,297,478 | IssuesEvent | 2021-05-21 07:45:02 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | The testcontainers are not able to pull the ryuk image | Source: Internal Team: Core Type: Test-Failure | Pulling the `ryuk` image fails in testcontainer tests. The issue is discussed here: https://github.com/testcontainers/testcontainers-java/issues/3574
It might be related to downgrading the testcontainers version in #18535. We could try if the new version 1.15.3 solves the problems.
**Stacktrace**
```
[INFO] Running com.hazelcast.jet.cdc.mysql.MySqlCdcIntegrationTest
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.994 s <<< FAILURE! - in com.hazelcast.jet.cdc.mysql.MySqlCdcIntegrationTest
[ERROR] customers(com.hazelcast.jet.cdc.mysql.MySqlCdcIntegrationTest) Time elapsed: 0.987 s <<< ERROR!
org.testcontainers.containers.ContainerLaunchException: Container startup failed
at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:330)
at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:311)
at org.testcontainers.containers.GenericContainer.starting(GenericContainer.java:1022)
at org.testcontainers.containers.FailureDetectingExternalResource$1.evaluate(FailureDetectingExternalResource.java:29)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at com.hazelcast.test.DumpBuildInfoOnFailureRule$1.evaluate(DumpBuildInfoOnFailureRule.java:37)
at com.hazelcast.test.metrics.MetricsRule$1.evaluate(MetricsRule.java:62)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at com.hazelcast.test.HazelcastSerialClassRunner.runChild(HazelcastSerialClassRunner.java:50)
at com.hazelcast.test.HazelcastSerialClassRunner.runChild(HazelcastSerialClassRunner.java:29)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at com.hazelcast.test.AbstractHazelcastClassRunner$1.evaluate(AbstractHazelcastClassRunner.java:306)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.testcontainers.containers.ContainerFetchException: Can't get Docker image: RemoteDockerImage(imageName=debezium/example-mysql:1.3, imagePullPolicy=DefaultPullPolicy())
at org.testcontainers.containers.GenericContainer.getDockerImageName(GenericContainer.java:1279)
at org.testcontainers.containers.GenericContainer.logger(GenericContainer.java:613)
at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:320)
... 22 more
Caused by: com.github.dockerjava.api.exception.NotFoundException: {"message":"No such image: testcontainersofficial/ryuk:0.3.0"}
at com.github.dockerjava.okhttp.OkHttpInvocationBuilder.execute(OkHttpInvocationBuilder.java:287)
at com.github.dockerjava.okhttp.OkHttpInvocationBuilder.execute(OkHttpInvocationBuilder.java:271)
at com.github.dockerjava.okhttp.OkHttpInvocationBuilder.post(OkHttpInvocationBuilder.java:129)
at com.github.dockerjava.core.exec.CreateContainerCmdExec.execute(CreateContainerCmdExec.java:33)
at com.github.dockerjava.core.exec.CreateContainerCmdExec.execute(CreateContainerCmdExec.java:13)
at com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21)
at com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35)
at com.github.dockerjava.core.command.CreateContainerCmdImpl.exec(CreateContainerCmdImpl.java:595)
at org.testcontainers.utility.ResourceReaper.start(ResourceReaper.java:94)
at org.testcontainers.DockerClientFactory.client(DockerClientFactory.java:168)
at org.testcontainers.LazyDockerClient.getDockerClient(LazyDockerClient.java:14)
at org.testcontainers.LazyDockerClient.listImagesCmd(LazyDockerClient.java:12)
at org.testcontainers.images.LocalImagesCache.maybeInitCache(LocalImagesCache.java:68)
at org.testcontainers.images.LocalImagesCache.get(LocalImagesCache.java:32)
at org.testcontainers.images.AbstractImagePullPolicy.shouldPull(AbstractImagePullPolicy.java:18)
at org.testcontainers.images.RemoteDockerImage.resolve(RemoteDockerImage.java:59)
at org.testcontainers.images.RemoteDockerImage.resolve(RemoteDockerImage.java:26)
at org.testcontainers.utility.LazyFuture.getResolvedValue(LazyFuture.java:20)
at org.testcontainers.utility.LazyFuture.get(LazyFuture.java:27)
at org.testcontainers.containers.GenericContainer.getDockerImageName(GenericContainer.java:1277)
... 24 more
``` | 1.0 | The testcontainers are not able to pull the ryuk image - Pulling the `ryuk` image fails in testcontainer tests. The issue is discussed here: https://github.com/testcontainers/testcontainers-java/issues/3574
It might be related to downgrading the testcontainers version in #18535. We could try if the new version 1.15.3 solves the problems.
**Stacktrace**
```
[INFO] Running com.hazelcast.jet.cdc.mysql.MySqlCdcIntegrationTest
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.994 s <<< FAILURE! - in com.hazelcast.jet.cdc.mysql.MySqlCdcIntegrationTest
[ERROR] customers(com.hazelcast.jet.cdc.mysql.MySqlCdcIntegrationTest) Time elapsed: 0.987 s <<< ERROR!
org.testcontainers.containers.ContainerLaunchException: Container startup failed
at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:330)
at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:311)
at org.testcontainers.containers.GenericContainer.starting(GenericContainer.java:1022)
at org.testcontainers.containers.FailureDetectingExternalResource$1.evaluate(FailureDetectingExternalResource.java:29)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at com.hazelcast.test.DumpBuildInfoOnFailureRule$1.evaluate(DumpBuildInfoOnFailureRule.java:37)
at com.hazelcast.test.metrics.MetricsRule$1.evaluate(MetricsRule.java:62)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at com.hazelcast.test.HazelcastSerialClassRunner.runChild(HazelcastSerialClassRunner.java:50)
at com.hazelcast.test.HazelcastSerialClassRunner.runChild(HazelcastSerialClassRunner.java:29)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at com.hazelcast.test.AbstractHazelcastClassRunner$1.evaluate(AbstractHazelcastClassRunner.java:306)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.testcontainers.containers.ContainerFetchException: Can't get Docker image: RemoteDockerImage(imageName=debezium/example-mysql:1.3, imagePullPolicy=DefaultPullPolicy())
at org.testcontainers.containers.GenericContainer.getDockerImageName(GenericContainer.java:1279)
at org.testcontainers.containers.GenericContainer.logger(GenericContainer.java:613)
at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:320)
... 22 more
Caused by: com.github.dockerjava.api.exception.NotFoundException: {"message":"No such image: testcontainersofficial/ryuk:0.3.0"}
at com.github.dockerjava.okhttp.OkHttpInvocationBuilder.execute(OkHttpInvocationBuilder.java:287)
at com.github.dockerjava.okhttp.OkHttpInvocationBuilder.execute(OkHttpInvocationBuilder.java:271)
at com.github.dockerjava.okhttp.OkHttpInvocationBuilder.post(OkHttpInvocationBuilder.java:129)
at com.github.dockerjava.core.exec.CreateContainerCmdExec.execute(CreateContainerCmdExec.java:33)
at com.github.dockerjava.core.exec.CreateContainerCmdExec.execute(CreateContainerCmdExec.java:13)
at com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21)
at com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35)
at com.github.dockerjava.core.command.CreateContainerCmdImpl.exec(CreateContainerCmdImpl.java:595)
at org.testcontainers.utility.ResourceReaper.start(ResourceReaper.java:94)
at org.testcontainers.DockerClientFactory.client(DockerClientFactory.java:168)
at org.testcontainers.LazyDockerClient.getDockerClient(LazyDockerClient.java:14)
at org.testcontainers.LazyDockerClient.listImagesCmd(LazyDockerClient.java:12)
at org.testcontainers.images.LocalImagesCache.maybeInitCache(LocalImagesCache.java:68)
at org.testcontainers.images.LocalImagesCache.get(LocalImagesCache.java:32)
at org.testcontainers.images.AbstractImagePullPolicy.shouldPull(AbstractImagePullPolicy.java:18)
at org.testcontainers.images.RemoteDockerImage.resolve(RemoteDockerImage.java:59)
at org.testcontainers.images.RemoteDockerImage.resolve(RemoteDockerImage.java:26)
at org.testcontainers.utility.LazyFuture.getResolvedValue(LazyFuture.java:20)
at org.testcontainers.utility.LazyFuture.get(LazyFuture.java:27)
at org.testcontainers.containers.GenericContainer.getDockerImageName(GenericContainer.java:1277)
... 24 more
``` | non_main | the testcontainers are not able to pull the ryuk image pulling the ryuk image fails in testcontainer tests the issue is discussed here it might be related to downgrading the testcontainers version in we could try if the new version solves the problems stacktrace running com hazelcast jet cdc mysql mysqlcdcintegrationtest tests run failures errors skipped time elapsed s failure in com hazelcast jet cdc mysql mysqlcdcintegrationtest customers com hazelcast jet cdc mysql mysqlcdcintegrationtest time elapsed s error org testcontainers containers containerlaunchexception container startup failed at org testcontainers containers genericcontainer dostart genericcontainer java at org testcontainers containers genericcontainer start genericcontainer java at org testcontainers containers genericcontainer starting genericcontainer java at org testcontainers containers failuredetectingexternalresource evaluate failuredetectingexternalresource java at org junit rules testwatcher evaluate testwatcher java at com hazelcast test dumpbuildinfoonfailurerule evaluate dumpbuildinfoonfailurerule java at com hazelcast test metrics metricsrule evaluate metricsrule java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at com hazelcast test hazelcastserialclassrunner runchild hazelcastserialclassrunner java at com hazelcast test hazelcastserialclassrunner runchild hazelcastserialclassrunner java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at com hazelcast test abstracthazelcastclassrunner evaluate abstracthazelcastclassrunner java at org junit internal runners statements failontimeout callablestatement call failontimeout java at org junit internal runners statements failontimeout callablestatement call failontimeout java at java util concurrent futuretask run futuretask java at java lang thread run thread java caused by org testcontainers containers containerfetchexception can t get docker image remotedockerimage imagename debezium example mysql imagepullpolicy defaultpullpolicy at org testcontainers containers genericcontainer getdockerimagename genericcontainer java at org testcontainers containers genericcontainer logger genericcontainer java at org testcontainers containers genericcontainer dostart genericcontainer java more caused by com github dockerjava api exception notfoundexception message no such image testcontainersofficial ryuk at com github dockerjava okhttp okhttpinvocationbuilder execute okhttpinvocationbuilder java at com github dockerjava okhttp okhttpinvocationbuilder execute okhttpinvocationbuilder java at com github dockerjava okhttp okhttpinvocationbuilder post okhttpinvocationbuilder java at com github dockerjava core exec createcontainercmdexec execute createcontainercmdexec java at com github dockerjava core exec createcontainercmdexec execute createcontainercmdexec java at com github dockerjava core exec abstrsyncdockercmdexec exec abstrsyncdockercmdexec java at com github dockerjava core command abstrdockercmd exec abstrdockercmd java at com github dockerjava core command createcontainercmdimpl exec createcontainercmdimpl java at org testcontainers utility resourcereaper start resourcereaper java at org testcontainers dockerclientfactory client dockerclientfactory java at org testcontainers lazydockerclient getdockerclient lazydockerclient java at org testcontainers lazydockerclient listimagescmd lazydockerclient java at org testcontainers images localimagescache maybeinitcache localimagescache java at org testcontainers images localimagescache get localimagescache java at org testcontainers images abstractimagepullpolicy shouldpull abstractimagepullpolicy java at org testcontainers images remotedockerimage resolve remotedockerimage java at org testcontainers images remotedockerimage resolve remotedockerimage java at org testcontainers utility lazyfuture getresolvedvalue lazyfuture java at org testcontainers utility lazyfuture get lazyfuture java at org testcontainers containers genericcontainer getdockerimagename genericcontainer java more | 0 |
5,562 | 27,822,775,132 | IssuesEvent | 2023-03-19 12:34:34 | shield-wall/myprofile | https://api.github.com/repos/shield-wall/myprofile | closed | Remove css from user website | enhancement layout maintainer | for now we are using many css to user's website [example](https://www.myprofile.pro/eerison)
the idea is use only [bulma](https://bulma.io) for this page. | True | Remove css from user website - for now we are using many css to user's website [example](https://www.myprofile.pro/eerison)
the idea is use only [bulma](https://bulma.io) for this page. | main | remove css from user website for now we are using many css to user s website the idea is use only for this page | 1 |
503,040 | 14,578,262,198 | IssuesEvent | 2020-12-18 04:20:09 | magento/magento2 | https://api.github.com/repos/magento/magento2 | opened | [GraphQl] Add wishlist item to cart Implementation | Priority: P1 Progress: ready for dev | Implement coverage for this new mutation
<https://github.com/magento/architecture/blob/master/design-documents/graph-ql/coverage/customer/Wishlist.graphqls]
`addWishlistItemsToCart( wishlistUid: ID!, @doc(description: "unique Id of wishlist") wishlistItemUids: [ID!> @doc(description: "Optional param. selected wish list items that are to be added") ): AddWishlistItemsToCartOutput @doc(description: "Add Requisition List Items To Customer Cart")`
### **`AC`**
* The field wishlistItemUids in the input can be null. Which implies that **all** the items in the wishlist will be added to the cart when wishlistItemUids are null.
* The **status** field in the response conveys the result of the operation true/false
* The add*wishlist_items_to_cart_user*errors field should aggregate all the errors occurred when adding the corresponding items to the cart.
| 1.0 | [GraphQl] Add wishlist item to cart Implementation - Implement coverage for this new mutation
<https://github.com/magento/architecture/blob/master/design-documents/graph-ql/coverage/customer/Wishlist.graphqls]
`addWishlistItemsToCart( wishlistUid: ID!, @doc(description: "unique Id of wishlist") wishlistItemUids: [ID!> @doc(description: "Optional param. selected wish list items that are to be added") ): AddWishlistItemsToCartOutput @doc(description: "Add Requisition List Items To Customer Cart")`
### **`AC`**
* The field wishlistItemUids in the input can be null. Which implies that **all** the items in the wishlist will be added to the cart when wishlistItemUids are null.
* The **status** field in the response conveys the result of the operation true/false
* The add*wishlist_items_to_cart_user*errors field should aggregate all the errors occurred when adding the corresponding items to the cart.
| non_main | add wishlist item to cart implementation implement coverage for this new mutation addwishlistitemstocart wishlistuid id doc description unique id of wishlist wishlistitemuids id doc description optional param selected wish list items that are to be added addwishlistitemstocartoutput doc description add requisition list items to customer cart ac the field wishlistitemuids in the input can be null which implies that all the items in the wishlist will be added to the cart when wishlistitemuids are null the status field in the response conveys the result of the operation true false the add wishlist items to cart user errors field should aggregate all the errors occurred when adding the corresponding items to the cart | 0 |
4,179 | 20,111,574,517 | IssuesEvent | 2022-02-07 15:31:29 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | opened | MiKo_3022 should not report for inherited methods | bug Area: analyzer Area: maintainability | As the return types cannot be changed for inherited methods, MiKo_3022 should not consider them. | True | MiKo_3022 should not report for inherited methods - As the return types cannot be changed for inherited methods, MiKo_3022 should not consider them. | main | miko should not report for inherited methods as the return types cannot be changed for inherited methods miko should not consider them | 1 |
43,948 | 5,719,686,767 | IssuesEvent | 2017-04-19 22:45:26 | brave/browser-laptop | https://api.github.com/repos/brave/browser-laptop | opened | Settings v Preferences mismatch | design feature/about-pages settings | - Did you search for similar issues before submitting this one?
Yes
- Describe the issue you encountered:
I was opening an issue for Sync, and noticed that we use _Settings_ from the top menu:
**_Edit > Settings_** for the `about:preferences` page.
Sync issue opened for reference: https://github.com/brave/sync/issues/79
- Platform (Win7, 8, 10? macOS? Linux distro?):
All (suspected) - need confirmation for macOS.
- Brave Version (revision SHA):
```
Brave: 0.14.2
rev: d4cad892de4c8da087c4efee2e8955caa630675c
Muon: 2.57.10
libchromiumcontent: 57.0.2987.133
V8: 5.7.492.71
Node.js: 7.4.0
Update Channel: dev
os.platform: win32
os.release: 10.0.14393
os.arch: x64
```
- Steps to reproduce:
Captured in this screenshot:

Within the Android browser, this is also referred to as Settings instead of Preferences. I'm not sure how much we can control for naming within the Android Brave browser, but I wanted to match if we're looking to have cross-platform consistency.
We're going to want to make sure we have consistency so the Sync instructions match across platforms.
@bradleyrichter I'm going to assign you on this once since we're talking about UI across platforms. Let me know if there's anything else I can assist with for this.
| 1.0 | Settings v Preferences mismatch - - Did you search for similar issues before submitting this one?
Yes
- Describe the issue you encountered:
I was opening an issue for Sync, and noticed that we use _Settings_ from the top menu:
**_Edit > Settings_** for the `about:preferences` page.
Sync issue opened for reference: https://github.com/brave/sync/issues/79
- Platform (Win7, 8, 10? macOS? Linux distro?):
All (suspected) - need confirmation for macOS.
- Brave Version (revision SHA):
```
Brave: 0.14.2
rev: d4cad892de4c8da087c4efee2e8955caa630675c
Muon: 2.57.10
libchromiumcontent: 57.0.2987.133
V8: 5.7.492.71
Node.js: 7.4.0
Update Channel: dev
os.platform: win32
os.release: 10.0.14393
os.arch: x64
```
- Steps to reproduce:
Captured in this screenshot:

Within the Android browser, this is also referred to as Settings instead of Preferences. I'm not sure how much we can control for naming within the Android Brave browser, but I wanted to match if we're looking to have cross-platform consistency.
We're going to want to make sure we have consistency so the Sync instructions match across platforms.
@bradleyrichter I'm going to assign you on this once since we're talking about UI across platforms. Let me know if there's anything else I can assist with for this.
| non_main | settings v preferences mismatch did you search for similar issues before submitting this one yes describe the issue you encountered i was opening an issue for sync and noticed that we use settings from the top menu edit settings for the about preferences page sync issue opened for reference platform macos linux distro all suspected need confirmation for macos brave version revision sha brave rev muon libchromiumcontent node js update channel dev os platform os release os arch steps to reproduce captured in this screenshot within the android browser this is also referred to as settings instead of preferences i m not sure how much we can control for naming within the android brave browser but i wanted to match if we re looking to have cross platform consistency we re going to want to make sure we have consistency so the sync instructions match across platforms bradleyrichter i m going to assign you on this once since we re talking about ui across platforms let me know if there s anything else i can assist with for this | 0 |
128,185 | 12,365,962,633 | IssuesEvent | 2020-05-18 09:39:56 | prometheus/docs | https://api.github.com/repos/prometheus/docs | closed | Visualization: Grafana Documentation Out-of-Date | wrong documentation | Following this documentation:
https://prometheus.io/docs/visualization/grafana/
The [configuring data source image](https://prometheus.io/assets/grafana_configuring_datasource.png) is out-of-date, and confusing given the [wording changes made to Direct / Proxy access](https://github.com/grafana/grafana/pull/11531).
Also, the documentation doesn't appear to be complete for a simple use case in docker.
Given
```
docker pull prom/prometheus
docker run -p 9090:9090 prom/prometheus
docker run -d --name=grafana -p 3000:3000 grafana/grafana
```
and using "Browser" access fails due to a CORS error I haven't quite figured out how to solve yet. Closest related issue is [this](https://github.com/grafana/grafana/issues/15225). The documentation should be updated with instructions for connecting when both prometheus and grafana are running as a docker container. | 1.0 | Visualization: Grafana Documentation Out-of-Date - Following this documentation:
https://prometheus.io/docs/visualization/grafana/
The [configuring data source image](https://prometheus.io/assets/grafana_configuring_datasource.png) is out-of-date, and confusing given the [wording changes made to Direct / Proxy access](https://github.com/grafana/grafana/pull/11531).
Also, the documentation doesn't appear to be complete for a simple use case in docker.
Given
```
docker pull prom/prometheus
docker run -p 9090:9090 prom/prometheus
docker run -d --name=grafana -p 3000:3000 grafana/grafana
```
and using "Browser" access fails due to a CORS error I haven't quite figured out how to solve yet. Closest related issue is [this](https://github.com/grafana/grafana/issues/15225). The documentation should be updated with instructions for connecting when both prometheus and grafana are running as a docker container. | non_main | visualization grafana documentation out of date following this documentation the is out of date and confusing given the also the documentation doesn t appear to be complete for a simple use case in docker given docker pull prom prometheus docker run p prom prometheus docker run d name grafana p grafana grafana and using browser access fails due to a cors error i haven t quite figured out how to solve yet closest related issue is the documentation should be updated with instructions for connecting when both prometheus and grafana are running as a docker container | 0 |
2,316 | 8,299,919,643 | IssuesEvent | 2018-09-21 05:59:32 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | reopened | Namespace names should not contain technical stuff such as 'Interface(s)', 'Exception(s)', 'ViewModel(s)', 'Implementation(s)' | Area: analyzer Area: maintainability Area: naming feature | Namespace names should be more functional than technical.
So they should not contains something like:
- Implementation
- Implementations
- Interface
- Interfaces
- Exception
- Exceptions
- Model
- Models
- Proxy
- Proxies
- ServiceProxy
- ServiceProxies
- View
- Views
- ViewModel
- ViewModels
In addtion, they should not contain useless stuff, such as
- Base
- Helper | True | Namespace names should not contain technical stuff such as 'Interface(s)', 'Exception(s)', 'ViewModel(s)', 'Implementation(s)' - Namespace names should be more functional than technical.
So they should not contains something like:
- Implementation
- Implementations
- Interface
- Interfaces
- Exception
- Exceptions
- Model
- Models
- Proxy
- Proxies
- ServiceProxy
- ServiceProxies
- View
- Views
- ViewModel
- ViewModels
In addtion, they should not contain useless stuff, such as
- Base
- Helper | main | namespace names should not contain technical stuff such as interface s exception s viewmodel s implementation s namespace names should be more functional than technical so they should not contains something like implementation implementations interface interfaces exception exceptions model models proxy proxies serviceproxy serviceproxies view views viewmodel viewmodels in addtion they should not contain useless stuff such as base helper | 1 |
3,804 | 16,456,001,301 | IssuesEvent | 2021-05-21 12:40:03 | cloverhearts/quilljs-markdown | https://api.github.com/repos/cloverhearts/quilljs-markdown | closed | [] is no style? | READY FOR MERGE Saw with Maintainer | target: [] => -[]
now: [] => dot
In addition, can you support tab input to indent? | True | [] is no style? - target: [] => -[]
now: [] => dot
In addition, can you support tab input to indent? | main | is no style target now dot in addition can you support tab input to indent? | 1 |
274,192 | 23,818,479,524 | IssuesEvent | 2022-09-05 09:02:46 | AhmedNSidd/gamers-social-manager-telegram-bot | https://api.github.com/repos/AhmedNSidd/gamers-social-manager-telegram-bot | closed | Prevent old invites from expiring | V1.0 Release user-testing discussion | Right now, our invite system needs some work to not automatically expire old invites if new ones are sent out. details will be discussed and shared here soon. jas's suggestion | 1.0 | Prevent old invites from expiring - Right now, our invite system needs some work to not automatically expire old invites if new ones are sent out. details will be discussed and shared here soon. jas's suggestion | non_main | prevent old invites from expiring right now our invite system needs some work to not automatically expire old invites if new ones are sent out details will be discussed and shared here soon jas s suggestion | 0 |
442,466 | 30,837,075,501 | IssuesEvent | 2023-08-02 08:09:23 | Kwonminwoo/board-fc | https://api.github.com/repos/Kwonminwoo/board-fc | closed | 깃헙 프로젝트와 이슈 정리하기 | documentation | 깃헙 프로젝트를 세팅하고, 카드를 만들어 정리
* [x] 프로젝트 베타 만들기
* [x] 카드 목록 만들기 - 강의 커리큘럼
* [x] 이슈로 적절히 바꾸기 | 1.0 | 깃헙 프로젝트와 이슈 정리하기 - 깃헙 프로젝트를 세팅하고, 카드를 만들어 정리
* [x] 프로젝트 베타 만들기
* [x] 카드 목록 만들기 - 강의 커리큘럼
* [x] 이슈로 적절히 바꾸기 | non_main | 깃헙 프로젝트와 이슈 정리하기 깃헙 프로젝트를 세팅하고 카드를 만들어 정리 프로젝트 베타 만들기 카드 목록 만들기 강의 커리큘럼 이슈로 적절히 바꾸기 | 0 |
359,072 | 10,659,689,323 | IssuesEvent | 2019-10-18 08:18:14 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.google.com - see bug description | ML Incorrect ML OFF browser-firefox-mobile engine-gecko priority-critical | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/search?q=mainboard+wechseln+ohne+Windows+neu&ie=utf-8&oe=utf-8&client=firefox-b-m
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: site freeze
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/10/ee8727f6-29f3-4b19-9b66-8951d9c6fed2.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191007162922</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.google.com - see bug description - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/search?q=mainboard+wechseln+ohne+Windows+neu&ie=utf-8&oe=utf-8&client=firefox-b-m
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: site freeze
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/10/ee8727f6-29f3-4b19-9b66-8951d9c6fed2.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191007162922</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description site freeze steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
3,361 | 13,031,367,547 | IssuesEvent | 2020-07-28 01:00:40 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | Maintainer change request - Content Access | Maintainer change request | Content Access is abandoned. I'm willing to take it over for now and try to get a full release:
https://github.com/backdrop-contrib/content_access/issues/5 | True | Maintainer change request - Content Access - Content Access is abandoned. I'm willing to take it over for now and try to get a full release:
https://github.com/backdrop-contrib/content_access/issues/5 | main | maintainer change request content access content access is abandoned i m willing to take it over for now and try to get a full release | 1 |
3,471 | 13,312,468,035 | IssuesEvent | 2020-08-26 09:46:43 | digitalpardoe/isyncit | https://api.github.com/repos/digitalpardoe/isyncit | closed | Power-Source-Based Bluetooth | Improvement No Longer Maintained | Add an option to turn bluetooth power on and off dependent on if the laptop is connected to a mains power source or not.
| True | Power-Source-Based Bluetooth - Add an option to turn bluetooth power on and off dependent on if the laptop is connected to a mains power source or not.
| main | power source based bluetooth add an option to turn bluetooth power on and off dependent on if the laptop is connected to a mains power source or not | 1 |
427,192 | 29,802,250,894 | IssuesEvent | 2023-06-16 08:59:46 | apecloud/kubeblocks | https://api.github.com/repos/apecloud/kubeblocks | closed | [Improvement] playground should check docker's version | kind/enhancement documentation | Is your improvement request related to a problem? Please describe.
`playground init` on local host better to check docker's version, if the version is too low, the k3d cluster will fail to create.
And better to output the operating system distro.
| 1.0 | [Improvement] playground should check docker's version - Is your improvement request related to a problem? Please describe.
`playground init` on local host better to check docker's version, if the version is too low, the k3d cluster will fail to create.
And better to output the operating system distro.
| non_main | playground should check docker s version is your improvement request related to a problem please describe playground init on local host better to check docker s version if the version is too low the cluster will fail to create and better to output the operating system distro | 0 |
32,528 | 13,878,856,448 | IssuesEvent | 2020-10-17 11:44:29 | microsoft/vscode-cpptools | https://api.github.com/repos/microsoft/vscode-cpptools | closed | Header autocomplete and header/source switch for "*.cu" "*.cuh" files | Feature Request Feature: Auto-complete Feature: Switch Header/Source Language Service more votes needed | Hi, I'm developing in Cuda right now and just adding `*.cu` and `*.cuh` to `files.associations` in settings works for syntax parsing.
However, when I try to autocomplete `#include <Header.cuh>`, it doesn't show files with `.cuh`/`.cu` extensions.
Also switching between `.cuh`/`.cu` as header/source doesn't work.
Is there any setting that can control this? | 1.0 | Header autocomplete and header/source switch for "*.cu" "*.cuh" files - Hi, I'm developing in Cuda right now and just adding `*.cu` and `*.cuh` to `files.associations` in settings works for syntax parsing.
However, when I try to autocomplete `#include <Header.cuh>`, it doesn't show files with `.cuh`/`.cu` extensions.
Also switching between `.cuh`/`.cu` as header/source doesn't work.
Is there any setting that can control this? | non_main | header autocomplete and header source switch for cu cuh files hi i m developing in cuda right now and just adding cu and cuh to files associations in settings works for syntax parsing however when i try to autocomplete include it doesn t show files with cuh cu extensions also switching between cuh cu as header source doesn t work is there any setting that can control this | 0 |
22,598 | 10,762,409,448 | IssuesEvent | 2019-10-31 23:34:40 | RuslanGox/BigRepo | https://api.github.com/repos/RuslanGox/BigRepo | opened | CVE-2018-1000620 (High) detected in cryptiles-2.0.5.tgz | security vulnerability | ## CVE-2018-1000620 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary>
<p>General purpose crypto utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/BigRepo/moodle-3.2.2/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/BigRepo/moodle-3.2.2/node_modules/cryptiles/package.json</p>
<p>
Dependency Hierarchy:
- grunt-contrib-less-1.3.0.tgz (Root Library)
- less-2.6.1.tgz
- request-2.73.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RuslanGox/BigRepo/commit/60f1b99636e9b000c240f2aa5f152ff189c55741">60f1b99636e9b000c240f2aa5f152ff189c55741</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-1000620">https://nvd.nist.gov/vuln/detail/CVE-2018-1000620</a></p>
<p>Release Date: 2019-04-08</p>
<p>Fix Resolution: 4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-1000620 (High) detected in cryptiles-2.0.5.tgz - ## CVE-2018-1000620 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary>
<p>General purpose crypto utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/BigRepo/moodle-3.2.2/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/BigRepo/moodle-3.2.2/node_modules/cryptiles/package.json</p>
<p>
Dependency Hierarchy:
- grunt-contrib-less-1.3.0.tgz (Root Library)
- less-2.6.1.tgz
- request-2.73.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RuslanGox/BigRepo/commit/60f1b99636e9b000c240f2aa5f152ff189c55741">60f1b99636e9b000c240f2aa5f152ff189c55741</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-1000620">https://nvd.nist.gov/vuln/detail/CVE-2018-1000620</a></p>
<p>Release Date: 2019-04-08</p>
<p>Fix Resolution: 4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in cryptiles tgz cve high severity vulnerability vulnerable library cryptiles tgz general purpose crypto utilities library home page a href path to dependency file tmp ws scm bigrepo moodle package json path to vulnerable library tmp ws scm bigrepo moodle node modules cryptiles package json dependency hierarchy grunt contrib less tgz root library less tgz request tgz hawk tgz x cryptiles tgz vulnerable library found in head commit a href vulnerability details eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
1,109 | 4,981,808,305 | IssuesEvent | 2016-12-07 09:19:44 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Things that are still del()'d | Maintainability - Hinders improvements - Not a bug | @duncathan
Here's a few things that are still `del`'d; not all of them should necessarily be `qdel`'d, but it's something to look at:
- some fake attacker images in hallucinations
- a lot of clients in the admin modules/topics
- inactivity
- SQL2 related things
- Client/New
I'm sure there's others.
| True | Things that are still del()'d - @duncathan
Here's a few things that are still `del`'d; not all of them should necessarily be `qdel`'d, but it's something to look at:
- some fake attacker images in hallucinations
- a lot of clients in the admin modules/topics
- inactivity
- SQL2 related things
- Client/New
I'm sure there's others.
| main | things that are still del d duncathan here s a few things that are still del d not all of them should necessarily be qdel d but it s something to look at some fake attacker images in hallucinations a lot of clients in the admin modules topics inactivity related things client new i m sure there s others | 1 |
4,163 | 19,978,730,903 | IssuesEvent | 2022-01-29 14:43:11 | TheOdinProject/curriculum | https://api.github.com/repos/TheOdinProject/curriculum | closed | Wrong Codepen on Intermediate HTML and CSS - Form Validation | Type: Bug Type: Maintainer Only | Escaping the backslash before a special character end up broke the form validation. All correct format is being evaluated as invalid

All codepens in the Pattern Validations section made this mistake. Backslash before special characters shouldn't be escaped as per this examples
https://www.w3schools.com/tags/att_input_pattern.asp | True | Wrong Codepen on Intermediate HTML and CSS - Form Validation - Escaping the backslash before a special character end up broke the form validation. All correct format is being evaluated as invalid

All codepens in the Pattern Validations section made this mistake. Backslash before special characters shouldn't be escaped as per this examples
https://www.w3schools.com/tags/att_input_pattern.asp | main | wrong codepen on intermediate html and css form validation escaping the backslash before a special character end up broke the form validation all correct format is being evaluated as invalid all codepens in the pattern validations section made this mistake backslash before special characters shouldn t be escaped as per this examples | 1 |
287,055 | 21,636,093,465 | IssuesEvent | 2022-05-05 14:21:01 | ethyca/fidesops | https://api.github.com/repos/ethyca/fidesops | closed | <masked_object_fields> docs example should include trailing comma or be at end of request body | documentation | ### Docs Update Description
A minor "bug" in saas config documents that tripped me up in some of my own testing.
Our `<masked_object_fields>` placeholder is replaced by a JSON serialization of fields that does include a trailing comma (`,`). This means the example provided leads users to a config that produces invalid JSON in their request at runtime.
The example should be updated to either include a `,` after the `<masked_object_fields>` placeholder, or the placeholder should be at the end of the request body (I prefer the former, as it gives more clarity on the behavior). Additionally, it may be good to provide a brief warning/info note about the fact that trailing commas are not included in the placeholder substitution, and therefore need to be included in the template.
old example:
```
body: '{
"properties": {
<masked_object_fields>
"user_ref_id": <user_ref_id>
}
}'
```
proposed new example:
```
body: '{
"properties": {
<masked_object_fields>,
"user_ref_id": <user_ref_id>
}
}'
```
(note the `,` following `<masked_object_fields>`)
### Additional context
One other potential "solution" I thought of was to update the placeholder substitution to include a trailing `,` after the fields, but I think that ultimately decreases flexibility or requires a bunch of string introspection that's probably not necessary. I think our current implementation is effective, so long as we've got this nuance clearly documented.
CC @ethyca/docs-authors
| 1.0 | <masked_object_fields> docs example should include trailing comma or be at end of request body - ### Docs Update Description
A minor "bug" in saas config documents that tripped me up in some of my own testing.
Our `<masked_object_fields>` placeholder is replaced by a JSON serialization of fields that does include a trailing comma (`,`). This means the example provided leads users to a config that produces invalid JSON in their request at runtime.
The example should be updated to either include a `,` after the `<masked_object_fields>` placeholder, or the placeholder should be at the end of the request body (I prefer the former, as it gives more clarity on the behavior). Additionally, it may be good to provide a brief warning/info note about the fact that trailing commas are not included in the placeholder substitution, and therefore need to be included in the template.
old example:
```
body: '{
"properties": {
<masked_object_fields>
"user_ref_id": <user_ref_id>
}
}'
```
proposed new example:
```
body: '{
"properties": {
<masked_object_fields>,
"user_ref_id": <user_ref_id>
}
}'
```
(note the `,` following `<masked_object_fields>`)
### Additional context
One other potential "solution" I thought of was to update the placeholder substitution to include a trailing `,` after the fields, but I think that ultimately decreases flexibility or requires a bunch of string introspection that's probably not necessary. I think our current implementation is effective, so long as we've got this nuance clearly documented.
CC @ethyca/docs-authors
| non_main | docs example should include trailing comma or be at end of request body docs update description a minor bug in saas config documents that tripped me up in some of my own testing our placeholder is replaced by a json serialization of fields that does include a trailing comma this means the example provided leads users to a config that produces invalid json in their request at runtime the example should be updated to either include a after the placeholder or the placeholder should be at the end of the request body i prefer the former as it gives more clarity on the behavior additionally it may be good to provide a brief warning info note about the fact that trailing commas are not included in the placeholder substitution and therefore need to be included in the template old example body properties user ref id proposed new example body properties user ref id note the following additional context one other potential solution i thought of was to update the placeholder substitution to include a trailing after the fields but i think that ultimately decreases flexibility or requires a bunch of string introspection that s probably not necessary i think our current implementation is effective so long as we ve got this nuance clearly documented cc ethyca docs authors | 0 |
230,010 | 7,603,291,505 | IssuesEvent | 2018-04-29 13:01:39 | alinaciuysal/OEDA | https://api.github.com/repos/alinaciuysal/OEDA | closed | results of the experiments | medium priority | - save experiment results (the whole result to be more generic) to the DB | 1.0 | results of the experiments - - save experiment results (the whole result to be more generic) to the DB | non_main | results of the experiments save experiment results the whole result to be more generic to the db | 0 |
571,962 | 17,023,393,085 | IssuesEvent | 2021-07-03 01:47:43 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | disable proxy for localhost | Component: merkaartor Priority: major Resolution: fixed Type: enhancement | **[Submitted to the original trac issue database at 9.56am, Monday, 27th April 2009]**
I have a proxy but use localhost for mapserver wms. So i have every time to go to preference to (un)check proxy settings (check to download or uplaod and uncheck to see wms images from mapserver).
but I think the best is to hoce a firefox-like setting (no proxy for multiple host) and/or automatic configuration script. | 1.0 | disable proxy for localhost - **[Submitted to the original trac issue database at 9.56am, Monday, 27th April 2009]**
I have a proxy but use localhost for mapserver wms. So i have every time to go to preference to (un)check proxy settings (check to download or uplaod and uncheck to see wms images from mapserver).
but I think the best is to hoce a firefox-like setting (no proxy for multiple host) and/or automatic configuration script. | non_main | disable proxy for localhost i have a proxy but use localhost for mapserver wms so i have every time to go to preference to un check proxy settings check to download or uplaod and uncheck to see wms images from mapserver but i think the best is to hoce a firefox like setting no proxy for multiple host and or automatic configuration script | 0 |
558,310 | 16,529,200,372 | IssuesEvent | 2021-05-27 02:04:28 | openmsupply/msupply-cold-chain | https://api.github.com/repos/openmsupply/msupply-cold-chain | closed | Can't update a sensors name | priority: high type: bug | **Describe the bug**
When trying to update a sensors name, it doesn't work!
**To Reproduce**
Steps to reproduce the behavior:
1. Go and try update a sensor name, apparently doesn't work!
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Reported by Arpita
| 1.0 | Can't update a sensors name - **Describe the bug**
When trying to update a sensors name, it doesn't work!
**To Reproduce**
Steps to reproduce the behavior:
1. Go and try update a sensor name, apparently doesn't work!
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Reported by Arpita
| non_main | can t update a sensors name describe the bug when trying to update a sensors name it doesn t work to reproduce steps to reproduce the behavior go and try update a sensor name apparently doesn t work see error expected behavior a clear and concise description of what you expected to happen smartphone please complete the following information device os browser version additional context reported by arpita | 0 |
241 | 2,970,850,340 | IssuesEvent | 2015-07-14 00:19:21 | cattolyst/datafinisher | https://api.github.com/repos/cattolyst/datafinisher | opened | Normalize the names of variables, files, and tables | maintainability | `dd2`,`data_dictionary`... these sorts of names will eventually cause confusion and maintainability issues. DataFinisher should have a consistent naming convention for the tables it creates, so they can be easily distinguished from native DataBuilder tables.
Also, `makesql.py` is the sort of thing that someone who has never seriously used Python with sqlite3 would use for a filename (and this is in fact the case, but why advertise it?). Should be renamed to `datafinisher.py`.
Also, to the extent that internal variable names don't consistently obey a naming convention, they should be fixed. | True | Normalize the names of variables, files, and tables - `dd2`,`data_dictionary`... these sorts of names will eventually cause confusion and maintainability issues. DataFinisher should have a consistent naming convention for the tables it creates, so they can be easily distinguished from native DataBuilder tables.
Also, `makesql.py` is the sort of thing that someone who has never seriously used Python with sqlite3 would use for a filename (and this is in fact the case, but why advertise it?). Should be renamed to `datafinisher.py`.
Also, to the extent that internal variable names don't consistently obey a naming convention, they should be fixed. | main | normalize the names of variables files and tables data dictionary these sorts of names will eventually cause confusion and maintainability issues datafinisher should have a consistent naming convention for the tables it creates so they can be easily distinguished from native databuilder tables also makesql py is the sort of thing that someone who has never seriously used python with would use for a filename and this is in fact the case but why advertise it should be renamed to datafinisher py also to the extent that internal variable names don t consistently obey a naming convention they should be fixed | 1 |
66,472 | 8,933,398,450 | IssuesEvent | 2019-01-23 01:51:13 | Sylius/Sylius | https://api.github.com/repos/Sylius/Sylius | closed | Custom cart contexts use-cases | Documentation Stale | **Sylius docs version**: latest
**Description**
In order to :
- allow users to fully enjoy the ability to add their own cart contexts,
- inform people of the customisation capabilities of Sylius,
The documentation should contain some e-commerce uses-cases about why and how one can add a customized cart context (implementing the `Sylius\Component\Order\Context\CartContextInterface`).
| 1.0 | Custom cart contexts use-cases - **Sylius docs version**: latest
**Description**
In order to :
- allow users to fully enjoy the ability to add their own cart contexts,
- inform people of the customisation capabilities of Sylius,
The documentation should contain some e-commerce uses-cases about why and how one can add a customized cart context (implementing the `Sylius\Component\Order\Context\CartContextInterface`).
| non_main | custom cart contexts use cases sylius docs version latest description in order to allow users to fully enjoy the ability to add their own cart contexts inform people of the customisation capabilities of sylius the documentation should contain some e commerce uses cases about why and how one can add a customized cart context implementing the sylius component order context cartcontextinterface | 0 |
194,149 | 15,397,968,649 | IssuesEvent | 2021-03-03 23:04:42 | yanniz0r/sonq | https://api.github.com/repos/yanniz0r/sonq | opened | Add readme | documentation | - [ ] Preview picture
- [ ] How to get the development started
- [ ] How to deploy this thing
- [ ] Contribution | 1.0 | Add readme - - [ ] Preview picture
- [ ] How to get the development started
- [ ] How to deploy this thing
- [ ] Contribution | non_main | add readme preview picture how to get the development started how to deploy this thing contribution | 0 |
536,784 | 15,713,274,392 | IssuesEvent | 2021-03-27 15:31:01 | AY2021S2-CS2103-T14-2/tp | https://api.github.com/repos/AY2021S2-CS2103-T14-2/tp | opened | Improved Price, Find & FindAll to include price as range | priority.Medium | As a user, I can search for prices by specifying a range, so that I can go for cheaper food options. | 1.0 | Improved Price, Find & FindAll to include price as range - As a user, I can search for prices by specifying a range, so that I can go for cheaper food options. | non_main | improved price find findall to include price as range as a user i can search for prices by specifying a range so that i can go for cheaper food options | 0 |
67,836 | 7,065,107,026 | IssuesEvent | 2018-01-06 16:03:18 | junit-team/junit5 | https://api.github.com/repos/junit-team/junit5 | closed | Cant create a nested parameterized test that uses @MethodSource | component: Jupiter status: waiting-for-feedback theme: parameterized tests | ## Overview
**Bug report.**
Version : org.junit.jupiter:junit-jupiter-params:5.0.2
Cant create a nested parameterized test that uses `@MethodSource`. `@MethodSource` requires that it points to a static method. But you cannot have a static method from a non-static inner class. And you cant have a nested class that is a static inner class
### Scenario 1
```java
import org.junit.jupiter.api.Nested;
import org.junit.jupiter.params.ParameterizedTest;
import org.junit.jupiter.params.provider.Arguments;
import org.junit.jupiter.params.provider.MethodSource;
import java.util.Arrays;
import java.util.stream.Stream;
import static org.junit.jupiter.api.Assertions.assertEquals;
class NestedParameterizedTest {
@Nested
class TheNesting {
@ParameterizedTest
@MethodSource("getTestData")
void shouldAdd(int a, int b, int c) {
assertEquals(c, a + b);
}
Stream<Arguments> getTestData() {
return Arrays.stream(new Arguments[]{
Arguments.of(1, 2, 3)
});
}
}
}
```
This code results to a `org.junit.platform.commons.util.PreconditionViolationException: Cannot invoke non-static method [java.util.stream.Stream<org.junit.jupiter.params.provider.Arguments> NestedParameterizedTest$TheNesting.getTestData()] on a null target.`
### Scenario 2
Making `getTestData()` static though results to a compile error `Inner classes cannot have static declarations`.
### Scenario 3
Making `getTestData()` and `class TheNesting` static though results in `Empty test suite.`
| 1.0 | Cant create a nested parameterized test that uses @MethodSource - ## Overview
**Bug report.**
Version : org.junit.jupiter:junit-jupiter-params:5.0.2
Cant create a nested parameterized test that uses `@MethodSource`. `@MethodSource` requires that it points to a static method. But you cannot have a static method from a non-static inner class. And you cant have a nested class that is a static inner class
### Scenario 1
```java
import org.junit.jupiter.api.Nested;
import org.junit.jupiter.params.ParameterizedTest;
import org.junit.jupiter.params.provider.Arguments;
import org.junit.jupiter.params.provider.MethodSource;
import java.util.Arrays;
import java.util.stream.Stream;
import static org.junit.jupiter.api.Assertions.assertEquals;
class NestedParameterizedTest {
@Nested
class TheNesting {
@ParameterizedTest
@MethodSource("getTestData")
void shouldAdd(int a, int b, int c) {
assertEquals(c, a + b);
}
Stream<Arguments> getTestData() {
return Arrays.stream(new Arguments[]{
Arguments.of(1, 2, 3)
});
}
}
}
```
This code results to a `org.junit.platform.commons.util.PreconditionViolationException: Cannot invoke non-static method [java.util.stream.Stream<org.junit.jupiter.params.provider.Arguments> NestedParameterizedTest$TheNesting.getTestData()] on a null target.`
### Scenario 2
Making `getTestData()` static though results to a compile error `Inner classes cannot have static declarations`.
### Scenario 3
Making `getTestData()` and `class TheNesting` static though results in `Empty test suite.`
| non_main | cant create a nested parameterized test that uses methodsource overview bug report version org junit jupiter junit jupiter params cant create a nested parameterized test that uses methodsource methodsource requires that it points to a static method but you cannot have a static method from a non static inner class and you cant have a nested class that is a static inner class scenario java import org junit jupiter api nested import org junit jupiter params parameterizedtest import org junit jupiter params provider arguments import org junit jupiter params provider methodsource import java util arrays import java util stream stream import static org junit jupiter api assertions assertequals class nestedparameterizedtest nested class thenesting parameterizedtest methodsource gettestdata void shouldadd int a int b int c assertequals c a b stream gettestdata return arrays stream new arguments arguments of this code results to a org junit platform commons util preconditionviolationexception cannot invoke non static method on a null target scenario making gettestdata static though results to a compile error inner classes cannot have static declarations scenario making gettestdata and class thenesting static though results in empty test suite | 0 |
5,538 | 27,716,568,265 | IssuesEvent | 2023-03-14 17:19:47 | cosmos/ibc-rs | https://api.github.com/repos/cosmos/ibc-rs | opened | [Crate] Take consistent approach for handling public structs/fields | O: maintainability | <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
Part of #25
Take a consistent approach all over the repo at places we have public structs/fields in terms of new instance creation or calling a specific field (e.g. in [`ics24_host::path`](https://github.com/cosmos/ibc-rs/blob/d5a44165c880b520bfb410ee693aa44ca8e6be7d/crates/ibc/src/core/ics24_host/path.rs#L48))
Should we either have public fields/structs or provide setters/getters/new/default methods (wherever required)?
| True | [Crate] Take consistent approach for handling public structs/fields - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
Part of #25
Take a consistent approach all over the repo at places we have public structs/fields in terms of new instance creation or calling a specific field (e.g. in [`ics24_host::path`](https://github.com/cosmos/ibc-rs/blob/d5a44165c880b520bfb410ee693aa44ca8e6be7d/crates/ibc/src/core/ics24_host/path.rs#L48))
Should we either have public fields/structs or provide setters/getters/new/default methods (wherever required)?
| main | take consistent approach for handling public structs fields ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ summary part of take a consistent approach all over the repo at places we have public structs fields in terms of new instance creation or calling a specific field e g in should we either have public fields structs or provide setters getters new default methods wherever required | 1 |
143,719 | 22,148,371,328 | IssuesEvent | 2022-06-03 14:18:28 | KeyWorksRW/wxUiEditor | https://api.github.com/repos/KeyWorksRW/wxUiEditor | opened | Remove prop_display_images from books | design change | ### Description:
<!-- Provide a description of of what you want to happen here -->
The various books that can contain images have a `display_images` property that controls whether or not images our displayed. wxUE uses that to determine whether or not to generate an image list for the book. It's a bit odd, though, for a page to specify a bitmap which then doesn't show up because the dev needs to know to turn it on in the book property.
I think we should remove this property, and instead rely completely on whether any of the pages specify a bitmap -- if they do, we create an image list and display the bitmap. | 1.0 | Remove prop_display_images from books - ### Description:
<!-- Provide a description of of what you want to happen here -->
The various books that can contain images have a `display_images` property that controls whether or not images our displayed. wxUE uses that to determine whether or not to generate an image list for the book. It's a bit odd, though, for a page to specify a bitmap which then doesn't show up because the dev needs to know to turn it on in the book property.
I think we should remove this property, and instead rely completely on whether any of the pages specify a bitmap -- if they do, we create an image list and display the bitmap. | non_main | remove prop display images from books description the various books that can contain images have a display images property that controls whether or not images our displayed wxue uses that to determine whether or not to generate an image list for the book it s a bit odd though for a page to specify a bitmap which then doesn t show up because the dev needs to know to turn it on in the book property i think we should remove this property and instead rely completely on whether any of the pages specify a bitmap if they do we create an image list and display the bitmap | 0 |
1,657 | 6,574,047,497 | IssuesEvent | 2017-09-11 11:14:40 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | apt-rpm Failed to get /usr/bin/rpm | affects_2.2 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apt-rpm
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
```
##### CONFIGURATION
```
$ cat ansible.cfg
[defaults]
transport = ssh
remote_user = toto
remote_port = 2222
host_key_checking = False
remote_tmp = /tmp
roles_path = ./roles
log_path = ./ansible.log
ansible_managed = Ansible managed: Don't modify manually modified on %Y-%m-%d %H:%M:%S
```
##### OS / ENVIRONMENT
`Debian stretch/sid`
##### SUMMARY
I can’t install a RPM located in roles/files/ with apt-rpm module
##### STEPS TO REPRODUCE
```
$ ll roles/icinga2/files/
total 16K
-rw-r--r-- 1 tr4sk tr4sk 4,3K juil. 12 12:21 icinga-rpm-release-6-1.el6.noarch.rpm
-rw-r--r-- 1 tr4sk tr4sk 4,3K juil. 12 12:25 icinga-rpm-release-7-1.el7.centos.noarch.rpm
- name: Add Icinga2 repo RPM
apt_rpm:
pkg: "icinga-rpm-release-{{ ansible_distribution_major_version }}-1.el{{ ansible_distribution_major_version }}.noarch.rpm"
state: present
TASK [icinga2 : Add Icinga2 repo RPM] ******************************************
fatal: [srv-sup.fr]: FAILED! => {"changed": false, "failed": true, "msg": "cannot find /usr/bin/apt-get and/or /usr/bin/rpm"}
```
On the remote host
```
# cat /etc/centos-release
CentOS release 6.8 (Final)
# whereis rpm
rpm: /bin/rpm /etc/rpm /usr/lib/rpm /usr/share/man/man8/rpm.8.gz
```
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
It shoulld install the pkg that I provided.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
TASK [icinga2 : Add Icinga2 repo RPM] ******************************************
task path: /home/tr4sk/script/toto/ansible-role-icinga2/tasks/RedHat_repo.yml:2
Using module file /home/tr4sk/.local/lib/python2.7/site-packages/ansible/modules/core/packaging/os/apt_rpm.py
<srv-sup.fr> ESTABLISH SSH CONNECTION FOR USER: toto
<srv-sup.fr> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r srv-sup.fr '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo /tmp/ansible-tmp-1480947022.71-23676108448873 `" && echo ansible-tmp-1480947022.71-23676108448873="` echo /tmp/ansible-tmp-1480947022.71-23676108448873 `" ) && sleep 0'"'"''
<srv-sup.fr> PUT /tmp/tmpvWStXj TO /tmp/ansible-tmp-1480947022.71-23676108448873/apt_rpm.py
<srv-sup.fr> SSH: EXEC sftp -b - -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r '[srv-sup.fr]'
<srv-sup.fr> ESTABLISH SSH CONNECTION FOR USER: toto
<srv-sup.fr> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r srv-sup.fr '/bin/sh -c '"'"'chmod u+x /tmp/ansible-tmp-1480947022.71-23676108448873/ /tmp/ansible-tmp-1480947022.71-23676108448873/apt_rpm.py && sleep 0'"'"''
<srv-sup.fr> ESTABLISH SSH CONNECTION FOR USER: toto
<srv-sup.fr> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r -tt srv-sup '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-iqfhuwvxfijpfxtptyamisppsnsfpmcl; /usr/bin/python /tmp/ansible-tmp-1480947022.71-23676108448873/apt_rpm.py; rm -rf "/tmp/ansible-tmp-1480947022.71-23676108448873/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
fatal: [srv-sup.fr]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"package": "icinga-rpm-release-6-1.el6.noarch.rpm",
"pkg": "icinga-rpm-release-6-1.el6.noarch.rpm",
"state": "present",
"update_cache": false
},
"module_name": "apt_rpm"
},
"msg": "cannot find /usr/bin/apt-get and/or /usr/bin/rpm"
}
```
| True | apt-rpm Failed to get /usr/bin/rpm - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apt-rpm
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
```
##### CONFIGURATION
```
$ cat ansible.cfg
[defaults]
transport = ssh
remote_user = toto
remote_port = 2222
host_key_checking = False
remote_tmp = /tmp
roles_path = ./roles
log_path = ./ansible.log
ansible_managed = Ansible managed: Don't modify manually modified on %Y-%m-%d %H:%M:%S
```
##### OS / ENVIRONMENT
`Debian stretch/sid`
##### SUMMARY
I can’t install a RPM located in roles/files/ with apt-rpm module
##### STEPS TO REPRODUCE
```
$ ll roles/icinga2/files/
total 16K
-rw-r--r-- 1 tr4sk tr4sk 4,3K juil. 12 12:21 icinga-rpm-release-6-1.el6.noarch.rpm
-rw-r--r-- 1 tr4sk tr4sk 4,3K juil. 12 12:25 icinga-rpm-release-7-1.el7.centos.noarch.rpm
- name: Add Icinga2 repo RPM
apt_rpm:
pkg: "icinga-rpm-release-{{ ansible_distribution_major_version }}-1.el{{ ansible_distribution_major_version }}.noarch.rpm"
state: present
TASK [icinga2 : Add Icinga2 repo RPM] ******************************************
fatal: [srv-sup.fr]: FAILED! => {"changed": false, "failed": true, "msg": "cannot find /usr/bin/apt-get and/or /usr/bin/rpm"}
```
On the remote host
```
# cat /etc/centos-release
CentOS release 6.8 (Final)
# whereis rpm
rpm: /bin/rpm /etc/rpm /usr/lib/rpm /usr/share/man/man8/rpm.8.gz
```
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
It shoulld install the pkg that I provided.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
TASK [icinga2 : Add Icinga2 repo RPM] ******************************************
task path: /home/tr4sk/script/toto/ansible-role-icinga2/tasks/RedHat_repo.yml:2
Using module file /home/tr4sk/.local/lib/python2.7/site-packages/ansible/modules/core/packaging/os/apt_rpm.py
<srv-sup.fr> ESTABLISH SSH CONNECTION FOR USER: toto
<srv-sup.fr> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r srv-sup.fr '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo /tmp/ansible-tmp-1480947022.71-23676108448873 `" && echo ansible-tmp-1480947022.71-23676108448873="` echo /tmp/ansible-tmp-1480947022.71-23676108448873 `" ) && sleep 0'"'"''
<srv-sup.fr> PUT /tmp/tmpvWStXj TO /tmp/ansible-tmp-1480947022.71-23676108448873/apt_rpm.py
<srv-sup.fr> SSH: EXEC sftp -b - -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r '[srv-sup.fr]'
<srv-sup.fr> ESTABLISH SSH CONNECTION FOR USER: toto
<srv-sup.fr> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r srv-sup.fr '/bin/sh -c '"'"'chmod u+x /tmp/ansible-tmp-1480947022.71-23676108448873/ /tmp/ansible-tmp-1480947022.71-23676108448873/apt_rpm.py && sleep 0'"'"''
<srv-sup.fr> ESTABLISH SSH CONNECTION FOR USER: toto
<srv-sup.fr> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2222 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=toto -o ConnectTimeout=10 -o ControlPath=/home/tr4sk/.ansible/cp/ansible-ssh-%h-%p-%r -tt srv-sup '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-iqfhuwvxfijpfxtptyamisppsnsfpmcl; /usr/bin/python /tmp/ansible-tmp-1480947022.71-23676108448873/apt_rpm.py; rm -rf "/tmp/ansible-tmp-1480947022.71-23676108448873/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
fatal: [srv-sup.fr]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"package": "icinga-rpm-release-6-1.el6.noarch.rpm",
"pkg": "icinga-rpm-release-6-1.el6.noarch.rpm",
"state": "present",
"update_cache": false
},
"module_name": "apt_rpm"
},
"msg": "cannot find /usr/bin/apt-get and/or /usr/bin/rpm"
}
```
| main | apt rpm failed to get usr bin rpm issue type bug report component name apt rpm ansible version ansible configuration cat ansible cfg transport ssh remote user toto remote port host key checking false remote tmp tmp roles path roles log path ansible log ansible managed ansible managed don t modify manually modified on y m d h m s os environment debian stretch sid summary i can’t install a rpm located in roles files with apt rpm module steps to reproduce ll roles files total rw r r juil icinga rpm release noarch rpm rw r r juil icinga rpm release centos noarch rpm name add repo rpm apt rpm pkg icinga rpm release ansible distribution major version el ansible distribution major version noarch rpm state present task fatal failed changed false failed true msg cannot find usr bin apt get and or usr bin rpm on the remote host cat etc centos release centos release final whereis rpm rpm bin rpm etc rpm usr lib rpm usr share man rpm gz expected results it shoulld install the pkg that i provided actual results task task path home script toto ansible role tasks redhat repo yml using module file home local lib site packages ansible modules core packaging os apt rpm py establish ssh connection for user toto ssh exec ssh vvv c o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user toto o connecttimeout o controlpath home ansible cp ansible ssh h p r srv sup fr bin sh c umask mkdir p echo tmp ansible tmp echo ansible tmp echo tmp ansible tmp sleep put tmp tmpvwstxj to tmp ansible tmp apt rpm py ssh exec sftp b vvv c o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user toto o connecttimeout o controlpath home ansible cp ansible ssh h p r establish ssh connection for user toto ssh exec ssh vvv c o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user toto o connecttimeout o controlpath home ansible cp ansible ssh h p r srv sup fr bin sh c chmod u x tmp ansible tmp tmp ansible tmp apt rpm py sleep establish ssh connection for user toto ssh exec ssh vvv c o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user toto o connecttimeout o controlpath home ansible cp ansible ssh h p r tt srv sup bin sh c sudo h s n u root bin sh c echo become success iqfhuwvxfijpfxtptyamisppsnsfpmcl usr bin python tmp ansible tmp apt rpm py rm rf tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module args package icinga rpm release noarch rpm pkg icinga rpm release noarch rpm state present update cache false module name apt rpm msg cannot find usr bin apt get and or usr bin rpm | 1 |
3,516 | 13,778,780,230 | IssuesEvent | 2020-10-08 12:56:58 | luakit/luakit | https://api.github.com/repos/luakit/luakit | closed | Gentoo ebuild needs update and maintainer | for-package-maintainer | https://packages.gentoo.org/packages/www-client/luakit
I see the text `This package needs a new maintainer!`. And available is only testing version for amd64, would be nice to have x86 also.

| True | Gentoo ebuild needs update and maintainer - https://packages.gentoo.org/packages/www-client/luakit
I see the text `This package needs a new maintainer!`. And available is only testing version for amd64, would be nice to have x86 also.

| main | gentoo ebuild needs update and maintainer i see the text this package needs a new maintainer and available is only testing version for would be nice to have also | 1 |
1,592 | 6,572,373,223 | IssuesEvent | 2017-09-11 01:48:41 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | known_hosts hashed hosts | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
- known_hosts
##### ANSIBLE VERSION
```
$ ansible --version
ansible 2.1.1.0
config file = (...)/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
None.
##### OS / ENVIRONMENT
- ubuntu/trusty64
##### SUMMARY
When using `known_hosts` module with hostname hash and `ssh-keyscan -H -t rsa example.com`,
on every execution, the hash will be different but the key type and key itself are the same.
This makes the module mark the task execution as `changed=True` even though the host, key type and key are the same.
##### STEPS TO REPRODUCE
For example (using shell commands to demo what is happening)...
First run:
```
# /etc/ssh/ssh_known_hosts is empty
$ ansible-playbook ...
- name: setup global known_hosts
known_hosts:
path=/etc/ssh/known_hosts
name="{{ item }}"
key="{{ lookup('pipe', 'ssh-keyscan -T 10 -H -t rsa ' + item) }}"
state=present
with_items:
- github.com
become: yes
# shell command was
$ sudo ssh-keyscan -H -t rsa github.com
# github.com SSH-2.0-libssh-0.7.0
|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk= ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAq2A7hRGmdnm9tUDbO9IDSwBK6TbQa+PXYPCPy6rbTrTtw7PHkccKrpp0yVhp5HdEIcKr6pLlVDBfOLX9QUsyCOV0wzfjIJNlGEYsdlLJizHhbn2mUjvSAHQqZETYP81eFzLQNnPHt4EVVUh7VfDESU84KezmD5QlWpXLmvU31/yMf+Se8xhHTvKSCZIFImWwoG6mbUoWf9nzpIoaSjB+weqqUUmpaaasXVal72J+UX2B+2RPW3RcT0eOzQgqlJL3RKrTJvdsjE3JEAvGq3lGHSZXy28G3skua2SmVi/w4yCE6gbODqnTWlg7+wC604ydGXA8VJiS5ap43JXiUFFAaQ==
# task marked as changed (good since there was no entry available)
```
The `/etc/ssh/ssh_known_hosts` file will then contain:
```
|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk= ssh-rsa AAAAB3NzaC1y
```
Second run
```
$ ansible-playbook ...
- name: setup global known_hosts
known_hosts:
path=/etc/ssh/known_hosts
name="{{ item }}"
key="{{ lookup('pipe', 'ssh-keyscan -T 10 -H -t rsa ' + item) }}"
state=present
with_items:
- github.com
become: yes
# shell command was
$ sudo ssh-keyscan -H -t rsa github.com
# github.com SSH-2.0-libssh-0.7.0
|1|y9vlIFjZwEbs0y9qnII52cdYRck=|q58dZ4+tDCSegG7EEuf8ToaDuvw= ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAq2A7hRGmdnm9tUDbO9IDSwBK6TbQa+PXYPCPy6rbTrTtw7PHkccKrpp0yVhp5HdEIcKr6pLlVDBfOLX9QUsyCOV0wzfjIJNlGEYsdlLJizHhbn2mUjvSAHQqZETYP81eFzLQNnPHt4EVVUh7VfDESU84KezmD5QlWpXLmvU31/yMf+Se8xhHTvKSCZIFImWwoG6mbUoWf9nzpIoaSjB+weqqUUmpaaasXVal72J+UX2B+2RPW3RcT0eOzQgqlJL3RKrTJvdsjE3JEAvGq3lGHSZXy28G3skua2SmVi/w4yCE6gbODqnTWlg7+wC604ydGXA8VJiS5ap43JXiUFFAaQ==
# task marked as changed!
```
The `/etc/ssh/ssh_known_hosts` file will then contain the new
```
|1|y9vlIFjZwEbs0y9qnII52cdYRck=|q58dZ4+tDCSegG7EEuf8ToaDuvw= ssh-rsa AAAAB3NzaC1y
```
and have the
```
|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk= ssh-rsa AAAAB3NzaC1y
```
removed.
It seems the module realizes that the `|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk=` is for the same host as `|1|y9vlIFjZwEbs0y9qnII52cdYRck=|q58dZ4+tDCSegG7EEuf8ToaDuvw=` or else it wouldn't remove it.
Since it's not adding a new key I don't think it should be marking the task as changed.
Not only that but it would be better if it didn't change the file if it didn't need to.
This is not a major issue but it does break any idempotency testing that uses this module and `ssh-keyscan -H` to setup known_hosts.
##### EXPECTED RESULTS
The module shouldn't mark the file as changed if the host hash changes but not the key type and key pair.
##### ACTUAL RESULTS
Every time I ran the `known_hosts` module and use the output of `ssh-keyscan -H -t rsa`,
the module is adding the new pair (host hash, type, key) and removing all of the previous hash for the same host.
| True | known_hosts hashed hosts - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
- known_hosts
##### ANSIBLE VERSION
```
$ ansible --version
ansible 2.1.1.0
config file = (...)/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
None.
##### OS / ENVIRONMENT
- ubuntu/trusty64
##### SUMMARY
When using `known_hosts` module with hostname hash and `ssh-keyscan -H -t rsa example.com`,
on every execution, the hash will be different but the key type and key itself are the same.
This makes the module mark the task execution as `changed=True` even though the host, key type and key are the same.
##### STEPS TO REPRODUCE
For example (using shell commands to demo what is happening)...
First run:
```
# /etc/ssh/ssh_known_hosts is empty
$ ansible-playbook ...
- name: setup global known_hosts
known_hosts:
path=/etc/ssh/known_hosts
name="{{ item }}"
key="{{ lookup('pipe', 'ssh-keyscan -T 10 -H -t rsa ' + item) }}"
state=present
with_items:
- github.com
become: yes
# shell command was
$ sudo ssh-keyscan -H -t rsa github.com
# github.com SSH-2.0-libssh-0.7.0
|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk= ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAq2A7hRGmdnm9tUDbO9IDSwBK6TbQa+PXYPCPy6rbTrTtw7PHkccKrpp0yVhp5HdEIcKr6pLlVDBfOLX9QUsyCOV0wzfjIJNlGEYsdlLJizHhbn2mUjvSAHQqZETYP81eFzLQNnPHt4EVVUh7VfDESU84KezmD5QlWpXLmvU31/yMf+Se8xhHTvKSCZIFImWwoG6mbUoWf9nzpIoaSjB+weqqUUmpaaasXVal72J+UX2B+2RPW3RcT0eOzQgqlJL3RKrTJvdsjE3JEAvGq3lGHSZXy28G3skua2SmVi/w4yCE6gbODqnTWlg7+wC604ydGXA8VJiS5ap43JXiUFFAaQ==
# task marked as changed (good since there was no entry available)
```
The `/etc/ssh/ssh_known_hosts` file will then contain:
```
|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk= ssh-rsa AAAAB3NzaC1y
```
Second run
```
$ ansible-playbook ...
- name: setup global known_hosts
known_hosts:
path=/etc/ssh/known_hosts
name="{{ item }}"
key="{{ lookup('pipe', 'ssh-keyscan -T 10 -H -t rsa ' + item) }}"
state=present
with_items:
- github.com
become: yes
# shell command was
$ sudo ssh-keyscan -H -t rsa github.com
# github.com SSH-2.0-libssh-0.7.0
|1|y9vlIFjZwEbs0y9qnII52cdYRck=|q58dZ4+tDCSegG7EEuf8ToaDuvw= ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAq2A7hRGmdnm9tUDbO9IDSwBK6TbQa+PXYPCPy6rbTrTtw7PHkccKrpp0yVhp5HdEIcKr6pLlVDBfOLX9QUsyCOV0wzfjIJNlGEYsdlLJizHhbn2mUjvSAHQqZETYP81eFzLQNnPHt4EVVUh7VfDESU84KezmD5QlWpXLmvU31/yMf+Se8xhHTvKSCZIFImWwoG6mbUoWf9nzpIoaSjB+weqqUUmpaaasXVal72J+UX2B+2RPW3RcT0eOzQgqlJL3RKrTJvdsjE3JEAvGq3lGHSZXy28G3skua2SmVi/w4yCE6gbODqnTWlg7+wC604ydGXA8VJiS5ap43JXiUFFAaQ==
# task marked as changed!
```
The `/etc/ssh/ssh_known_hosts` file will then contain the new
```
|1|y9vlIFjZwEbs0y9qnII52cdYRck=|q58dZ4+tDCSegG7EEuf8ToaDuvw= ssh-rsa AAAAB3NzaC1y
```
and have the
```
|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk= ssh-rsa AAAAB3NzaC1y
```
removed.
It seems the module realizes that the `|1|4lzJnCDEsWzp/zt0polmlEUesFk=|m/y3U10yA03Vsy3uvvVpZo+t/Gk=` is for the same host as `|1|y9vlIFjZwEbs0y9qnII52cdYRck=|q58dZ4+tDCSegG7EEuf8ToaDuvw=` or else it wouldn't remove it.
Since it's not adding a new key I don't think it should be marking the task as changed.
Not only that but it would be better if it didn't change the file if it didn't need to.
This is not a major issue but it does break any idempotency testing that uses this module and `ssh-keyscan -H` to setup known_hosts.
##### EXPECTED RESULTS
The module shouldn't mark the file as changed if the host hash changes but not the key type and key pair.
##### ACTUAL RESULTS
Every time I ran the `known_hosts` module and use the output of `ssh-keyscan -H -t rsa`,
the module is adding the new pair (host hash, type, key) and removing all of the previous hash for the same host.
| main | known hosts hashed hosts issue type bug report component name known hosts ansible version ansible version ansible config file ansible cfg configured module search path default w o overrides configuration none os environment ubuntu summary when using known hosts module with hostname hash and ssh keyscan h t rsa example com on every execution the hash will be different but the key type and key itself are the same this makes the module mark the task execution as changed true even though the host key type and key are the same steps to reproduce for example using shell commands to demo what is happening first run etc ssh ssh known hosts is empty ansible playbook name setup global known hosts known hosts path etc ssh known hosts name item key lookup pipe ssh keyscan t h t rsa item state present with items github com become yes shell command was sudo ssh keyscan h t rsa github com github com ssh libssh m t gk ssh rsa ymf task marked as changed good since there was no entry available the etc ssh ssh known hosts file will then contain m t gk ssh rsa second run ansible playbook name setup global known hosts known hosts path etc ssh known hosts name item key lookup pipe ssh keyscan t h t rsa item state present with items github com become yes shell command was sudo ssh keyscan h t rsa github com github com ssh libssh ssh rsa ymf task marked as changed the etc ssh ssh known hosts file will then contain the new ssh rsa and have the m t gk ssh rsa removed it seems the module realizes that the m t gk is for the same host as or else it wouldn t remove it since it s not adding a new key i don t think it should be marking the task as changed not only that but it would be better if it didn t change the file if it didn t need to this is not a major issue but it does break any idempotency testing that uses this module and ssh keyscan h to setup known hosts expected results the module shouldn t mark the file as changed if the host hash changes but not the key type and key pair actual results every time i ran the known hosts module and use the output of ssh keyscan h t rsa the module is adding the new pair host hash type key and removing all of the previous hash for the same host | 1 |
1,708 | 6,574,437,078 | IssuesEvent | 2017-09-11 12:53:49 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | docker_image : make action explicit | affects_2.2 cloud docker feature_idea waiting_on_maintainer | Hi,
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
docker_images
##### ANSIBLE VERSION
```
root$ ansible --version
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = ['/usr/share/ansible']
```
##### CONFIGURATION
Not relevant
##### OS / ENVIRONMENT
Not relevant
##### SUMMARY
The docker_images module is very difficult to use because there is no action attribute to tells what we want to do. The action is implicit (if path -> build, if tag sometimes it tags sometimes not, ...). It would be more easy to make the action explicit with an action attribute: tag -> tags (alias) an image, build -> build, ...
##### STEPS TO REPRODUCE
Example of difficulty to use there: http://stackoverflow.com/questions/38169244/how-do-i-tag-a-local-docker-image-with-ansible-docker-image-module
How to do the equivalent of:
```
docker tag <imageSrc> <imageDest> ?
```
| True | docker_image : make action explicit - Hi,
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
docker_images
##### ANSIBLE VERSION
```
root$ ansible --version
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = ['/usr/share/ansible']
```
##### CONFIGURATION
Not relevant
##### OS / ENVIRONMENT
Not relevant
##### SUMMARY
The docker_images module is very difficult to use because there is no action attribute to tells what we want to do. The action is implicit (if path -> build, if tag sometimes it tags sometimes not, ...). It would be more easy to make the action explicit with an action attribute: tag -> tags (alias) an image, build -> build, ...
##### STEPS TO REPRODUCE
Example of difficulty to use there: http://stackoverflow.com/questions/38169244/how-do-i-tag-a-local-docker-image-with-ansible-docker-image-module
How to do the equivalent of:
```
docker tag <imageSrc> <imageDest> ?
```
| main | docker image make action explicit hi issue type feature idea component name docker images ansible version root ansible version ansible config file etc ansible ansible cfg configured module search path configuration not relevant os environment not relevant summary the docker images module is very difficult to use because there is no action attribute to tells what we want to do the action is implicit if path build if tag sometimes it tags sometimes not it would be more easy to make the action explicit with an action attribute tag tags alias an image build build steps to reproduce example of difficulty to use there how to do the equivalent of docker tag | 1 |
29,967 | 5,967,786,914 | IssuesEvent | 2017-05-30 16:40:46 | buildo/react-components | https://api.github.com/repos/buildo/react-components | opened | React peerDependency is still <0.15 | defect | ## description
When installing brc on a React 15 project npm yields
```
├─┬ buildo-react-components@0.22.1
│ └── UNMET PEER DEPENDENCY react@>=0.13.0 <0.15.0 || ^0.14.0-beta3
└── UNMET PEER DEPENDENCY eslint-plugin-jsx-a11y@5.0.1
```
## specs
update the peerDep | 1.0 | React peerDependency is still <0.15 - ## description
When installing brc on a React 15 project npm yields
```
├─┬ buildo-react-components@0.22.1
│ └── UNMET PEER DEPENDENCY react@>=0.13.0 <0.15.0 || ^0.14.0-beta3
└── UNMET PEER DEPENDENCY eslint-plugin-jsx-a11y@5.0.1
```
## specs
update the peerDep | non_main | react peerdependency is still description when installing brc on a react project npm yields ├─┬ buildo react components │ └── unmet peer dependency react └── unmet peer dependency eslint plugin jsx specs update the peerdep | 0 |
4,871 | 25,020,356,394 | IssuesEvent | 2022-11-03 23:32:38 | deislabs/spiderlightning | https://api.github.com/repos/deislabs/spiderlightning | closed | #18 epic: allow user to "watch" for events/changes | ⭐ epic 🚧 maintainer issue ❎ not active | @devigned and @khenidak discussed we want some interface for the event-sourcing, or change data capture( CDC) pattern.
```go
events.
Listen([kv.On("somefilter"), sql.On("some other filter")]).
Exec(|event| {
code...
})
```
```go
kv:= get_kv(some_name)
kv.On('someFilter', 'some description of events').Exec(kvInterface, k, v){
code...
}
```
There will be a series of PRs to tackle on this problem.
- [x] #51
- [x] #50
- [x] #66
- [x] #65
- [x] #140
- [x] #141
- [x] #68
- [ ] #110
- [ ] #111
- [ ] #112
- [ ] #113
- [x] #63
- [x] #53 | True | #18 epic: allow user to "watch" for events/changes - @devigned and @khenidak discussed we want some interface for the event-sourcing, or change data capture( CDC) pattern.
```go
events.
Listen([kv.On("somefilter"), sql.On("some other filter")]).
Exec(|event| {
code...
})
```
```go
kv:= get_kv(some_name)
kv.On('someFilter', 'some description of events').Exec(kvInterface, k, v){
code...
}
```
There will be a series of PRs to tackle on this problem.
- [x] #51
- [x] #50
- [x] #66
- [x] #65
- [x] #140
- [x] #141
- [x] #68
- [ ] #110
- [ ] #111
- [ ] #112
- [ ] #113
- [x] #63
- [x] #53 | main | epic allow user to watch for events changes devigned and khenidak discussed we want some interface for the event sourcing or change data capture cdc pattern go events listen exec event code go kv get kv some name kv on somefilter some description of events exec kvinterface k v code there will be a series of prs to tackle on this problem | 1 |
137,597 | 12,760,256,266 | IssuesEvent | 2020-06-29 07:42:04 | bradharms/degu | https://api.github.com/repos/bradharms/degu | opened | Installation Process | documentation enhancement | Define and document a process by which a first-time user would install the software. | 1.0 | Installation Process - Define and document a process by which a first-time user would install the software. | non_main | installation process define and document a process by which a first time user would install the software | 0 |
101,831 | 16,528,747,207 | IssuesEvent | 2021-05-27 01:05:03 | RG4421/azure-iot-platform-dotnet | https://api.github.com/repos/RG4421/azure-iot-platform-dotnet | opened | CVE-2021-32640 (Medium) detected in ws-6.2.1.tgz | security vulnerability | ## CVE-2021-32640 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ws-6.2.1.tgz</b></p></summary>
<p>Simple to use, blazing fast and thoroughly tested websocket client and server for Node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/ws/-/ws-6.2.1.tgz">https://registry.npmjs.org/ws/-/ws-6.2.1.tgz</a></p>
<p>Path to dependency file: azure-iot-platform-dotnet/src/webui/azure-iot-ux-fluent-controls/package.json</p>
<p>Path to vulnerable library: azure-iot-platform-dotnet/src/webui/azure-iot-ux-fluent-controls/node_modules/ws/package.json</p>
<p>
Dependency Hierarchy:
- jsdom-12.2.0.tgz (Root Library)
- :x: **ws-6.2.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ws is an open source WebSocket client and server library for Node.js. A specially crafted value of the `Sec-Websocket-Protocol` header can be used to significantly slow down a ws server. The vulnerability has been fixed in ws@7.4.6 (https://github.com/websockets/ws/commit/00c425ec77993773d823f018f64a5c44e17023ff). In vulnerable versions of ws, the issue can be mitigated by reducing the maximum allowed length of the request headers using the [`--max-http-header-size=size`](https://nodejs.org/api/cli.html#cli_max_http_header_size_size) and/or the [`maxHeaderSize`](https://nodejs.org/api/http.html#http_http_createserver_options_requestlistener) options.
<p>Publish Date: 2021-05-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32640>CVE-2021-32640</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693">https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693</a></p>
<p>Release Date: 2021-05-25</p>
<p>Fix Resolution: ws - 7.4.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ws","packageVersion":"6.2.1","packageFilePaths":["/src/webui/azure-iot-ux-fluent-controls/package.json"],"isTransitiveDependency":true,"dependencyTree":"jsdom:12.2.0;ws:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ws - 7.4.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32640","vulnerabilityDetails":"ws is an open source WebSocket client and server library for Node.js. A specially crafted value of the `Sec-Websocket-Protocol` header can be used to significantly slow down a ws server. The vulnerability has been fixed in ws@7.4.6 (https://github.com/websockets/ws/commit/00c425ec77993773d823f018f64a5c44e17023ff). In vulnerable versions of ws, the issue can be mitigated by reducing the maximum allowed length of the request headers using the [`--max-http-header-size\u003dsize`](https://nodejs.org/api/cli.html#cli_max_http_header_size_size) and/or the [`maxHeaderSize`](https://nodejs.org/api/http.html#http_http_createserver_options_requestlistener) options.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32640","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-32640 (Medium) detected in ws-6.2.1.tgz - ## CVE-2021-32640 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ws-6.2.1.tgz</b></p></summary>
<p>Simple to use, blazing fast and thoroughly tested websocket client and server for Node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/ws/-/ws-6.2.1.tgz">https://registry.npmjs.org/ws/-/ws-6.2.1.tgz</a></p>
<p>Path to dependency file: azure-iot-platform-dotnet/src/webui/azure-iot-ux-fluent-controls/package.json</p>
<p>Path to vulnerable library: azure-iot-platform-dotnet/src/webui/azure-iot-ux-fluent-controls/node_modules/ws/package.json</p>
<p>
Dependency Hierarchy:
- jsdom-12.2.0.tgz (Root Library)
- :x: **ws-6.2.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ws is an open source WebSocket client and server library for Node.js. A specially crafted value of the `Sec-Websocket-Protocol` header can be used to significantly slow down a ws server. The vulnerability has been fixed in ws@7.4.6 (https://github.com/websockets/ws/commit/00c425ec77993773d823f018f64a5c44e17023ff). In vulnerable versions of ws, the issue can be mitigated by reducing the maximum allowed length of the request headers using the [`--max-http-header-size=size`](https://nodejs.org/api/cli.html#cli_max_http_header_size_size) and/or the [`maxHeaderSize`](https://nodejs.org/api/http.html#http_http_createserver_options_requestlistener) options.
<p>Publish Date: 2021-05-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32640>CVE-2021-32640</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693">https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693</a></p>
<p>Release Date: 2021-05-25</p>
<p>Fix Resolution: ws - 7.4.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ws","packageVersion":"6.2.1","packageFilePaths":["/src/webui/azure-iot-ux-fluent-controls/package.json"],"isTransitiveDependency":true,"dependencyTree":"jsdom:12.2.0;ws:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ws - 7.4.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32640","vulnerabilityDetails":"ws is an open source WebSocket client and server library for Node.js. A specially crafted value of the `Sec-Websocket-Protocol` header can be used to significantly slow down a ws server. The vulnerability has been fixed in ws@7.4.6 (https://github.com/websockets/ws/commit/00c425ec77993773d823f018f64a5c44e17023ff). In vulnerable versions of ws, the issue can be mitigated by reducing the maximum allowed length of the request headers using the [`--max-http-header-size\u003dsize`](https://nodejs.org/api/cli.html#cli_max_http_header_size_size) and/or the [`maxHeaderSize`](https://nodejs.org/api/http.html#http_http_createserver_options_requestlistener) options.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32640","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_main | cve medium detected in ws tgz cve medium severity vulnerability vulnerable library ws tgz simple to use blazing fast and thoroughly tested websocket client and server for node js library home page a href path to dependency file azure iot platform dotnet src webui azure iot ux fluent controls package json path to vulnerable library azure iot platform dotnet src webui azure iot ux fluent controls node modules ws package json dependency hierarchy jsdom tgz root library x ws tgz vulnerable library found in base branch master vulnerability details ws is an open source websocket client and server library for node js a specially crafted value of the sec websocket protocol header can be used to significantly slow down a ws server the vulnerability has been fixed in ws in vulnerable versions of ws the issue can be mitigated by reducing the maximum allowed length of the request headers using the and or the options publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ws isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree jsdom ws isminimumfixversionavailable true minimumfixversion ws basebranches vulnerabilityidentifier cve vulnerabilitydetails ws is an open source websocket client and server library for node js a specially crafted value of the sec websocket protocol header can be used to significantly slow down a ws server the vulnerability has been fixed in ws in vulnerable versions of ws the issue can be mitigated by reducing the maximum allowed length of the request headers using the and or the options vulnerabilityurl | 0 |
4,562 | 23,729,382,251 | IssuesEvent | 2022-08-30 23:23:53 | deislabs/spiderlightning | https://api.github.com/repos/deislabs/spiderlightning | reopened | make an observability interface | 📐 proposal 🚧 maintainer issue | We don't have an interface that allows observability like Dapr does, but we definitely need one. This task will involve the design of an interface, the implementation of a host that exports it, an example that imports it, and testing. | True | make an observability interface - We don't have an interface that allows observability like Dapr does, but we definitely need one. This task will involve the design of an interface, the implementation of a host that exports it, an example that imports it, and testing. | main | make an observability interface we don t have an interface that allows observability like dapr does but we definitely need one this task will involve the design of an interface the implementation of a host that exports it an example that imports it and testing | 1 |
229,625 | 18,395,809,734 | IssuesEvent | 2021-10-12 11:12:38 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | Notebook Editor: showNotebookDocment integration test failing | integration-test-failure | https://dev.azure.com/monacotools/Monaco/_build/results?buildId=129832&view=logs&j=4d2898ab-dfbe-557e-92e7-aaac158fdd2f&t=b2ac4860-fb3b-5a51-42a7-7423ead0721b
```
Notebook Editor
showNotebookDocment:
AssertionError [ERR_ASSERTION] [ERR_ASSERTION]: Expected values to be strictly equal:
+ actual - expected
+ 'untitled:Untitled-2.nbdserializer?notebook.nbdserializer'
- 'fake-fs:/pdnkfv.nbdtest'
+ expected - actual
-untitled:Untitled-2.nbdserializer?notebook.nbdserializer
+fake-fs:/pdnkfv.nbdtest
at Context.<anonymous> (extensions/vscode-api-tests/src/singlefolder-tests/notebook.editor.test.ts:56:10)
``` | 1.0 | Notebook Editor: showNotebookDocment integration test failing - https://dev.azure.com/monacotools/Monaco/_build/results?buildId=129832&view=logs&j=4d2898ab-dfbe-557e-92e7-aaac158fdd2f&t=b2ac4860-fb3b-5a51-42a7-7423ead0721b
```
Notebook Editor
showNotebookDocment:
AssertionError [ERR_ASSERTION] [ERR_ASSERTION]: Expected values to be strictly equal:
+ actual - expected
+ 'untitled:Untitled-2.nbdserializer?notebook.nbdserializer'
- 'fake-fs:/pdnkfv.nbdtest'
+ expected - actual
-untitled:Untitled-2.nbdserializer?notebook.nbdserializer
+fake-fs:/pdnkfv.nbdtest
at Context.<anonymous> (extensions/vscode-api-tests/src/singlefolder-tests/notebook.editor.test.ts:56:10)
``` | non_main | notebook editor shownotebookdocment integration test failing notebook editor shownotebookdocment assertionerror expected values to be strictly equal actual expected untitled untitled nbdserializer notebook nbdserializer fake fs pdnkfv nbdtest expected actual untitled untitled nbdserializer notebook nbdserializer fake fs pdnkfv nbdtest at context extensions vscode api tests src singlefolder tests notebook editor test ts | 0 |
5,369 | 26,996,968,838 | IssuesEvent | 2023-02-10 02:27:26 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | sam init quickstart templates failing on windows | platform/windows area/init maintainer/need-followup | ### Description:
```
sam init {{ any flags }}
```
will fail with fresh installs on windows. I can replicate with defaults or a smattering of various options. (For reference, I'm trying to create a node16.x project Hello World project).
I have SAM and AWS CLI installed on a fresh windows 11 machine. I can replicate the issue with identical output with a fresh install of both CLIs on a windows 10 machine. The CLI will make it through all interactive prompts or flags properly, but throw an error when actually creating the project locally. I can repro the problem in command prompt or Powershell (PSCore 7.x) .The non-verbose error is
```
Error: Unstable state when updating repo. Check that you have permissions to create/delete files in C:\Users\taran\AppData\Roaming\AWS SAM directory or file an issue at https://github.com/aws/aws-sam-cli/issues
```
### Steps to reproduce:
- Start with clean (No AWS products installed) install of Windows 11 (the issue seems to effect windows 10 too so dealers choice).
- Ensure python and git are installed
- Install latest AWS CLI and SAM CLI from their respective MSIs, pulled from the getting started docs.
- Create a fresh AWS account, create a user with the generic Admin role. Create CLI credentials for this user. (not sure much of this matters but this is what I did) (Enable MFA on the root user, do more best practices etc etc etc).
- Run aws configure using the aforementioned access keys
- create a blank directory and cd into it
- run sam init -> quickstart templates -> option 10 for node 16 -> option 1 for zip -> option 1 for not typescript. -> No to Xray -> default app name
(Not sure if it makes a difference but I did have docker desktop installed and running, I tried this with image deployments too. Same results)
### Observed result:
```
PS C:\Users\taran\SAM> sam init --debug
2022-07-02 16:50:13,858 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2022-07-02 16:50:13,858 | Using config file: samconfig.toml, config environment: default
2022-07-02 16:50:13,858 | Expand command line arguments to:
2022-07-02 16:50:13,858 | --package_type=Zip --output_dir=.
You can preselect a particular runtime or package type when using the `sam init` experience.
Call `sam init --help` to learn more.
Which template source would you like to use?
1 - AWS Quick Start Templates
2 - Custom Template Location
Choice: 1
Choose an AWS Quick Start application template
1 - Hello World Example
2 - Multi-step workflow
3 - Serverless API
4 - Scheduled task
5 - Standalone function
6 - Data processing
7 - Infrastructure event management
8 - Machine Learning
Template: 1
Use the most popular runtime and package type? (Python and zip) [y/N]: N
Which runtime would you like to use?
1 - dotnet6
2 - dotnet5.0
3 - dotnetcore3.1
4 - go1.x
5 - graalvm.java11 (provided.al2)
6 - graalvm.java17 (provided.al2)
7 - java11
8 - java8.al2
9 - java8
10 - nodejs16.x
11 - nodejs14.x
12 - nodejs12.x
13 - python3.9
14 - python3.8
15 - python3.7
16 - python3.6
17 - ruby2.7
18 - rust (provided.al2)
Runtime: 10
What package type would you like to use?
1 - Zip
2 - Image
Package type: 1
Based on your selections, the only dependency manager available is npm.
We will proceed copying the template using npm.
Select your starter template
1 - Hello World Example
2 - Hello World Example TypeScript
Template: 1
Would you like to enable X-Ray tracing on the function(s) in your application? [y/N]: N
Project name [sam-app]:
2022-07-02 16:50:32,915 |
Cloning from https://github.com/aws/aws-sam-cli-app-templates (process may take a moment)
2022-07-02 16:50:32,916 | Configure core.longpaths=true in git clone. You might also need to enable long paths in Windows registry.
2022-07-02 16:50:44,289 | Removing old repo at C:\Users\taran\AppData\Roaming\AWS SAM\aws-sam-cli-app-templates
2022-07-02 16:50:45,685 | Copying from C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates to C:\Users\taran\AppData\Roaming\AWS SAM\aws-sam-cli-app-templates
2022-07-02 16:50:52,140 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,140 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,140 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,141 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,141 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,141 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,142 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,142 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,143 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,144 | rmtree failed in <built-in function unlink> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld\native-image.properties, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219900>)
2022-07-02 16:50:52,146 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219900>)
2022-07-02 16:50:52,147 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2198C0>)
2022-07-02 16:50:52,148 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,149 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2197C0>)
2022-07-02 16:50:52,150 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219780>)
2022-07-02 16:50:52,152 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219600>)
2022-07-02 16:50:52,153 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2195C0>)
2022-07-02 16:50:52,156 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219500>)
2022-07-02 16:50:52,157 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2192C0>)
2022-07-02 16:50:52,164 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,166 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,166 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,166 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,168 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,170 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219740>)
2022-07-02 16:50:52,171 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2196C0>)
2022-07-02 16:50:52,172 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219780>)
2022-07-02 16:50:52,173 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219640>)
2022-07-02 16:50:52,177 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2194C0>)
2022-07-02 16:50:52,179 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219480>)
2022-07-02 16:50:52,182 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219380>)
2022-07-02 16:50:52,183 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219300>)
2022-07-02 16:50:52,185 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005EC0>)
2022-07-02 16:50:52,194 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,195 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,195 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,197 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,198 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,198 | rmtree failed in <built-in function unlink> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld\native-image.properties, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219840>)
2022-07-02 16:50:52,200 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219840>)
2022-07-02 16:50:52,201 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219900>)
2022-07-02 16:50:52,202 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,202 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219700>)
2022-07-02 16:50:52,203 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219740>)
2022-07-02 16:50:52,206 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219780>)
2022-07-02 16:50:52,207 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219640>)
2022-07-02 16:50:52,209 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219440>)
2022-07-02 16:50:52,210 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219200>)
2022-07-02 16:50:52,218 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,219 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,220 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,220 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,220 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,221 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,221 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,221 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,222 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,224 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2196C0>)
2022-07-02 16:50:52,225 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219680>)
2022-07-02 16:50:52,226 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219740>)
2022-07-02 16:50:52,227 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219540>)
2022-07-02 16:50:52,230 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2194C0>)
2022-07-02 16:50:52,231 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219400>)
2022-07-02 16:50:52,233 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219380>)
2022-07-02 16:50:52,234 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2191C0>)
2022-07-02 16:50:52,236 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005EC0>)
2022-07-02 16:50:52,237 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005F00>)
2022-07-02 16:50:52,241 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005F40>)
2022-07-02 16:50:52,563 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005E00>)
2022-07-02 16:50:52,564 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E69FE8100>)
2022-07-02 16:50:52,565 | Template is not provided in context, skip adding project type metric
2022-07-02 16:50:52,566 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': '4051ee11-2e95-423d-ab51-2d657db33725', 'installationId': 'eda45204-2efc-44cd-b956-ac7563effaad', 'sessionId': '5840cdc0-ce22-4ebb-a35a-18419c48e119', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.8', 'samcliVersion': '1.53.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam init', 'duration': 38708, 'exitReason': 'AppTemplateUpdateException', 'exitCode': 1}}]}
2022-07-02 16:50:53,150 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Error: Unstable state when updating repo. Check that you have permissions to create/delete files in C:\Users\taran\AppData\Roaming\AWS SAM directory or file an issue at https://github.com/aws/aws-sam-cli/issues
PS C:\Users\taran\SAM>
```
### Expected result:
No errors, fresh SAM template fun
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: WIndows 11 OS Build 22000.739
2. `sam --version`: 1.53.0
3. AWS region: us-east-1
AWS CLI version: 2.7.12
git for windows 2.37.0 | True | sam init quickstart templates failing on windows - ### Description:
```
sam init {{ any flags }}
```
will fail with fresh installs on windows. I can replicate with defaults or a smattering of various options. (For reference, I'm trying to create a node16.x project Hello World project).
I have SAM and AWS CLI installed on a fresh windows 11 machine. I can replicate the issue with identical output with a fresh install of both CLIs on a windows 10 machine. The CLI will make it through all interactive prompts or flags properly, but throw an error when actually creating the project locally. I can repro the problem in command prompt or Powershell (PSCore 7.x) .The non-verbose error is
```
Error: Unstable state when updating repo. Check that you have permissions to create/delete files in C:\Users\taran\AppData\Roaming\AWS SAM directory or file an issue at https://github.com/aws/aws-sam-cli/issues
```
### Steps to reproduce:
- Start with clean (No AWS products installed) install of Windows 11 (the issue seems to effect windows 10 too so dealers choice).
- Ensure python and git are installed
- Install latest AWS CLI and SAM CLI from their respective MSIs, pulled from the getting started docs.
- Create a fresh AWS account, create a user with the generic Admin role. Create CLI credentials for this user. (not sure much of this matters but this is what I did) (Enable MFA on the root user, do more best practices etc etc etc).
- Run aws configure using the aforementioned access keys
- create a blank directory and cd into it
- run sam init -> quickstart templates -> option 10 for node 16 -> option 1 for zip -> option 1 for not typescript. -> No to Xray -> default app name
(Not sure if it makes a difference but I did have docker desktop installed and running, I tried this with image deployments too. Same results)
### Observed result:
```
PS C:\Users\taran\SAM> sam init --debug
2022-07-02 16:50:13,858 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2022-07-02 16:50:13,858 | Using config file: samconfig.toml, config environment: default
2022-07-02 16:50:13,858 | Expand command line arguments to:
2022-07-02 16:50:13,858 | --package_type=Zip --output_dir=.
You can preselect a particular runtime or package type when using the `sam init` experience.
Call `sam init --help` to learn more.
Which template source would you like to use?
1 - AWS Quick Start Templates
2 - Custom Template Location
Choice: 1
Choose an AWS Quick Start application template
1 - Hello World Example
2 - Multi-step workflow
3 - Serverless API
4 - Scheduled task
5 - Standalone function
6 - Data processing
7 - Infrastructure event management
8 - Machine Learning
Template: 1
Use the most popular runtime and package type? (Python and zip) [y/N]: N
Which runtime would you like to use?
1 - dotnet6
2 - dotnet5.0
3 - dotnetcore3.1
4 - go1.x
5 - graalvm.java11 (provided.al2)
6 - graalvm.java17 (provided.al2)
7 - java11
8 - java8.al2
9 - java8
10 - nodejs16.x
11 - nodejs14.x
12 - nodejs12.x
13 - python3.9
14 - python3.8
15 - python3.7
16 - python3.6
17 - ruby2.7
18 - rust (provided.al2)
Runtime: 10
What package type would you like to use?
1 - Zip
2 - Image
Package type: 1
Based on your selections, the only dependency manager available is npm.
We will proceed copying the template using npm.
Select your starter template
1 - Hello World Example
2 - Hello World Example TypeScript
Template: 1
Would you like to enable X-Ray tracing on the function(s) in your application? [y/N]: N
Project name [sam-app]:
2022-07-02 16:50:32,915 |
Cloning from https://github.com/aws/aws-sam-cli-app-templates (process may take a moment)
2022-07-02 16:50:32,916 | Configure core.longpaths=true in git clone. You might also need to enable long paths in Windows registry.
2022-07-02 16:50:44,289 | Removing old repo at C:\Users\taran\AppData\Roaming\AWS SAM\aws-sam-cli-app-templates
2022-07-02 16:50:45,685 | Copying from C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates to C:\Users\taran\AppData\Roaming\AWS SAM\aws-sam-cli-app-templates
2022-07-02 16:50:52,140 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,140 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,140 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,141 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,141 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,141 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,142 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,142 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,143 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219940>)
2022-07-02 16:50:52,144 | rmtree failed in <built-in function unlink> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld\native-image.properties, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219900>)
2022-07-02 16:50:52,146 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219900>)
2022-07-02 16:50:52,147 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2198C0>)
2022-07-02 16:50:52,148 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,149 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2197C0>)
2022-07-02 16:50:52,150 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219780>)
2022-07-02 16:50:52,152 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219600>)
2022-07-02 16:50:52,153 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2195C0>)
2022-07-02 16:50:52,156 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219500>)
2022-07-02 16:50:52,157 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-gradle, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2192C0>)
2022-07-02 16:50:52,164 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,166 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,166 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,166 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,167 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,168 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,170 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219740>)
2022-07-02 16:50:52,171 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2196C0>)
2022-07-02 16:50:52,172 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219780>)
2022-07-02 16:50:52,173 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219640>)
2022-07-02 16:50:52,177 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2194C0>)
2022-07-02 16:50:52,179 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219480>)
2022-07-02 16:50:52,182 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219380>)
2022-07-02 16:50:52,183 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11\cookiecutter-aws-sam-graalvm-maven, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219300>)
2022-07-02 16:50:52,185 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java11, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005EC0>)
2022-07-02 16:50:52,194 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,195 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,195 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,196 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,197 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,198 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219980>)
2022-07-02 16:50:52,198 | rmtree failed in <built-in function unlink> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld\native-image.properties, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219840>)
2022-07-02 16:50:52,200 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\helloworld, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219840>)
2022-07-02 16:50:52,201 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219900>)
2022-07-02 16:50:52,202 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,202 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219700>)
2022-07-02 16:50:52,203 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219740>)
2022-07-02 16:50:52,206 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219780>)
2022-07-02 16:50:52,207 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219640>)
2022-07-02 16:50:52,209 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219440>)
2022-07-02 16:50:52,210 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-gradle, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219200>)
2022-07-02 16:50:52,218 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,219 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-core, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,220 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,220 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-events, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,220 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,221 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-runtime-interface-client, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,221 | rmtree failed in <built-in function scandir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,221 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws\aws-lambda-java-serialization, details: (<class 'FileNotFoundError'>, FileNotFoundError(2, 'The system cannot find the path specified'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,222 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image\com.amazonaws, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219880>)
2022-07-02 16:50:52,224 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF\native-image, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2196C0>)
2022-07-02 16:50:52,225 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources\META-INF, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219680>)
2022-07-02 16:50:52,226 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main\resources, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219740>)
2022-07-02 16:50:52,227 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src\main, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219540>)
2022-07-02 16:50:52,230 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction\src, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2194C0>)
2022-07-02 16:50:52,231 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}\HelloWorldFunction, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219400>)
2022-07-02 16:50:52,233 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven\{{cookiecutter.project_name}}, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A219380>)
2022-07-02 16:50:52,234 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17\cookiecutter-aws-sam-graalvm-maven, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A2191C0>)
2022-07-02 16:50:52,236 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm\java17, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005EC0>)
2022-07-02 16:50:52,237 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2\graalvm, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005F00>)
2022-07-02 16:50:52,241 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates\provided.al2, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005F40>)
2022-07-02 16:50:52,563 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls\aws-sam-cli-app-templates, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E6A005E00>)
2022-07-02 16:50:52,564 | rmtree failed in <built-in function rmdir> for C:\Users\taran\AppData\Local\Temp\tmpkn0ngyls, details: (<class 'OSError'>, OSError(41, 'The directory is not empty'), <traceback object at 0x0000021E69FE8100>)
2022-07-02 16:50:52,565 | Template is not provided in context, skip adding project type metric
2022-07-02 16:50:52,566 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': '4051ee11-2e95-423d-ab51-2d657db33725', 'installationId': 'eda45204-2efc-44cd-b956-ac7563effaad', 'sessionId': '5840cdc0-ce22-4ebb-a35a-18419c48e119', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.8', 'samcliVersion': '1.53.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam init', 'duration': 38708, 'exitReason': 'AppTemplateUpdateException', 'exitCode': 1}}]}
2022-07-02 16:50:53,150 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Error: Unstable state when updating repo. Check that you have permissions to create/delete files in C:\Users\taran\AppData\Roaming\AWS SAM directory or file an issue at https://github.com/aws/aws-sam-cli/issues
PS C:\Users\taran\SAM>
```
### Expected result:
No errors, fresh SAM template fun
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: WIndows 11 OS Build 22000.739
2. `sam --version`: 1.53.0
3. AWS region: us-east-1
AWS CLI version: 2.7.12
git for windows 2.37.0 | main | sam init quickstart templates failing on windows description sam init any flags will fail with fresh installs on windows i can replicate with defaults or a smattering of various options for reference i m trying to create a x project hello world project i have sam and aws cli installed on a fresh windows machine i can replicate the issue with identical output with a fresh install of both clis on a windows machine the cli will make it through all interactive prompts or flags properly but throw an error when actually creating the project locally i can repro the problem in command prompt or powershell pscore x the non verbose error is error unstable state when updating repo check that you have permissions to create delete files in c users taran appdata roaming aws sam directory or file an issue at steps to reproduce start with clean no aws products installed install of windows the issue seems to effect windows too so dealers choice ensure python and git are installed install latest aws cli and sam cli from their respective msis pulled from the getting started docs create a fresh aws account create a user with the generic admin role create cli credentials for this user not sure much of this matters but this is what i did enable mfa on the root user do more best practices etc etc etc run aws configure using the aforementioned access keys create a blank directory and cd into it run sam init quickstart templates option for node option for zip option for not typescript no to xray default app name not sure if it makes a difference but i did have docker desktop installed and running i tried this with image deployments too same results observed result ps c users taran sam sam init debug telemetry endpoint configured to be using config file samconfig toml config environment default expand command line arguments to package type zip output dir you can preselect a particular runtime or package type when using the sam init experience call sam init help to learn more which template source would you like to use aws quick start templates custom template location choice choose an aws quick start application template hello world example multi step workflow serverless api scheduled task standalone function data processing infrastructure event management machine learning template use the most popular runtime and package type python and zip n which runtime would you like to use x graalvm provided graalvm provided x x x rust provided runtime what package type would you like to use zip image package type based on your selections the only dependency manager available is npm we will proceed copying the template using npm select your starter template hello world example hello world example typescript template would you like to enable x ray tracing on the function s in your application n project name cloning from process may take a moment configure core longpaths true in git clone you might also need to enable long paths in windows registry removing old repo at c users taran appdata roaming aws sam aws sam cli app templates copying from c users taran appdata local temp aws sam cli app templates to c users taran appdata roaming aws sam aws sam cli app templates rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image helloworld native image properties details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image helloworld details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image helloworld native image properties details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image helloworld details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf native image details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources meta inf details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main resources details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src main details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction src details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name helloworldfunction details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle cookiecutter project name details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm gradle details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java core details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java events details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java runtime interface client details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws aws lambda java serialization details filenotfounderror the system cannot find the path specified rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image com amazonaws details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf native image details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources meta inf details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main resources details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src main details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction src details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name helloworldfunction details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven cookiecutter project name details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm cookiecutter aws sam graalvm maven details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided graalvm details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates provided details oserror the directory is not empty rmtree failed in for c users taran appdata local temp aws sam cli app templates details oserror the directory is not empty rmtree failed in for c users taran appdata local temp details oserror the directory is not empty template is not provided in context skip adding project type metric sending telemetry metrics httpsconnectionpool host aws serverless tools telemetry us west amazonaws com port read timed out read timeout error unstable state when updating repo check that you have permissions to create delete files in c users taran appdata roaming aws sam directory or file an issue at ps c users taran sam expected result no errors fresh sam template fun additional environment details ex windows mac amazon linux etc os windows os build sam version aws region us east aws cli version git for windows | 1 |
2,084 | 7,093,728,424 | IssuesEvent | 2018-01-12 21:48:52 | ocaml/opam-repository | https://api.github.com/repos/ocaml/opam-repository | closed | CAMLCity is down. OPAM's URLs are timing out for ocamlfind | needs maintainer action | The package download URL listed here does not work and it appears that camlcity.org (which hosts all the findlib tarballs) is down entirely:
https://opam.ocaml.org/packages/ocamlfind/ocamlfind.1.7.3/ | True | CAMLCity is down. OPAM's URLs are timing out for ocamlfind - The package download URL listed here does not work and it appears that camlcity.org (which hosts all the findlib tarballs) is down entirely:
https://opam.ocaml.org/packages/ocamlfind/ocamlfind.1.7.3/ | main | camlcity is down opam s urls are timing out for ocamlfind the package download url listed here does not work and it appears that camlcity org which hosts all the findlib tarballs is down entirely | 1 |
5,241 | 26,563,281,160 | IssuesEvent | 2023-01-20 17:40:10 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Parse user entry of dates on the front end | type: enhancement work: frontend status: ready restricted: maintainers | ## Context
In the following places, we accept user-entry of date/datetime/time values:
- Cells, when saving values
- Filter conditions, when filtering the table page, or an exploration
- Inputs when searching via the record selector
- Inputs when saving a record on the Table Page
- Inputs, when setting a default value for a column
## Current behavior
- In all the above contexts, the front end attempts to parse the user input into a structured value. If it succeeds, it sends an ISO-formatted string for that value. If it fails, it sends the raw user input. Sending the raw user input allows us to handle special values like "today" and "now" that postgres is able to understand. We don't want to parse those values into dates on the front end because we want to allow the user to set filter conditions that will change based on the current date (e.g. setting the default value for a date column to "today").
- This behavior is problematic because it produces this bug described in #1890.

## Desired behavior
- The front end maintains a list of reserved "special" date values, so far only `today`, `now`, `yesterday`. If the user input matches a special value, then the front end sends that special value. Otherwise...
- The front end attempts to parse the user input into a structured value. If it succeeds, it sends an ISO-formatted string for that value. If it fails, then it displays a client-side error, allowing the user to fix the problem before submitting any API requests.
## Additional considerations
- I noticed that for some formats (e.g. `1/1/2020`) the Postgres parsing succeeds but our current parsing logic on the front end does not. When implementing this ticket, we'll need to improve the parsing logic on the front end to handle more cases like this and avoid UX regressions.
| True | Parse user entry of dates on the front end - ## Context
In the following places, we accept user-entry of date/datetime/time values:
- Cells, when saving values
- Filter conditions, when filtering the table page, or an exploration
- Inputs when searching via the record selector
- Inputs when saving a record on the Table Page
- Inputs, when setting a default value for a column
## Current behavior
- In all the above contexts, the front end attempts to parse the user input into a structured value. If it succeeds, it sends an ISO-formatted string for that value. If it fails, it sends the raw user input. Sending the raw user input allows us to handle special values like "today" and "now" that postgres is able to understand. We don't want to parse those values into dates on the front end because we want to allow the user to set filter conditions that will change based on the current date (e.g. setting the default value for a date column to "today").
- This behavior is problematic because it produces this bug described in #1890.

## Desired behavior
- The front end maintains a list of reserved "special" date values, so far only `today`, `now`, `yesterday`. If the user input matches a special value, then the front end sends that special value. Otherwise...
- The front end attempts to parse the user input into a structured value. If it succeeds, it sends an ISO-formatted string for that value. If it fails, then it displays a client-side error, allowing the user to fix the problem before submitting any API requests.
## Additional considerations
- I noticed that for some formats (e.g. `1/1/2020`) the Postgres parsing succeeds but our current parsing logic on the front end does not. When implementing this ticket, we'll need to improve the parsing logic on the front end to handle more cases like this and avoid UX regressions.
| main | parse user entry of dates on the front end context in the following places we accept user entry of date datetime time values cells when saving values filter conditions when filtering the table page or an exploration inputs when searching via the record selector inputs when saving a record on the table page inputs when setting a default value for a column current behavior in all the above contexts the front end attempts to parse the user input into a structured value if it succeeds it sends an iso formatted string for that value if it fails it sends the raw user input sending the raw user input allows us to handle special values like today and now that postgres is able to understand we don t want to parse those values into dates on the front end because we want to allow the user to set filter conditions that will change based on the current date e g setting the default value for a date column to today this behavior is problematic because it produces this bug described in desired behavior the front end maintains a list of reserved special date values so far only today now yesterday if the user input matches a special value then the front end sends that special value otherwise the front end attempts to parse the user input into a structured value if it succeeds it sends an iso formatted string for that value if it fails then it displays a client side error allowing the user to fix the problem before submitting any api requests additional considerations i noticed that for some formats e g the postgres parsing succeeds but our current parsing logic on the front end does not when implementing this ticket we ll need to improve the parsing logic on the front end to handle more cases like this and avoid ux regressions | 1 |
5,025 | 25,787,223,785 | IssuesEvent | 2022-12-09 21:56:35 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Auth problem with the records endpoint after session expiry | type: bug work: backend status: ready restricted: maintainers | - After my session on the staging site expired, I pressed the "Refresh" UI button within the Table Page.
- The columns and constraints endpoints responded appropriately with an HTTP 403 and the following body:
```json
[
{
"code": 4003,
"message": "Authentication credentials were not provided.",
"details": { "exception": "Authentication credentials were not provided." }
}
]
```
- But the records endpoint gave me an HTTP 500 and the following Django error
<details>
<summary>Traceback</summary>
```
Environment:
Request Method: GET
Request URL: https://staging.mathesar.org/api/db/v0/tables/12454/records/?limit=500&offset=0&order_by=%5B%7B%22field%22%3A79900%2C%22direction%22%3A%22asc%22%7D%5D&filter=%7B%22and%22%3A%5B%7B%22equal%22%3A%5B%7B%22column_id%22%3A%5B62229%5D%7D%2C%7B%22literal%22%3A%5B8%5D%7D%5D%7D%2C%7B%22lesser%22%3A%5B%7B%22column_id%22%3A%5B62697%5D%7D%2C%7B%22literal%22%3A%5B%22100%22%5D%7D%5D%7D%5D%7D
Django Version: 3.1.14
Python Version: 3.9.2
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
'mathesar.middleware.CursorClosedHandlerMiddleware',
'mathesar.middleware.PasswordChangeNeededMiddleware',
'django_userforeignkey.middleware.UserForeignKeyMiddleware',
'django_request_cache.middleware.RequestCacheMiddleware']
Traceback (most recent call last):
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/fields/__init__.py", line 1774, in get_prep_value
return int(value)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/contrib/auth/models.py", line 414, in __int__
raise TypeError('Cannot cast AnonymousUser to int. Are you trying to use it in place of User?')
The above exception (Cannot cast AnonymousUser to int. Are you trying to use it in place of User?) was the direct cause of the following exception:
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/var/www/staging.mathesar.org/mathesar/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler
raise exc
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 497, in dispatch
self.initial(request, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 415, in initial
self.check_permissions(request)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 332, in check_permissions
if not permission.has_permission(request, self):
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 53, in has_permission
allowed = self._evaluate_statements(statements, request, view, action)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 101, in _evaluate_statements
matched = self._get_statements_matching_conditions(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 235, in _get_statements_matching_conditions
passed = bool(boolExpr.parseString(condition)[0])
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/parsing.py", line 32, in __bool__
return self.evalop(bool(a) for a in self.args)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/parsing.py", line 32, in <genexpr>
return self.evalop(bool(a) for a in self.args)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/parsing.py", line 14, in __bool__
return self.check_condition_fn(self.label)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 218, in <lambda>
check_cond_fn = lambda cond: self._check_condition(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 263, in _check_condition
result = method(request, view, action)
File "/var/www/staging.mathesar.org/mathesar/mathesar/api/db/permissions/records.py", line 37, in is_table_viewer
is_schema_viewer = SchemaRole.objects.filter(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/query.py", line 942, in filter
return self._filter_or_exclude(False, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/query.py", line 962, in _filter_or_exclude
clone._filter_or_exclude_inplace(negate, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/query.py", line 969, in _filter_or_exclude_inplace
self._query.add_q(Q(*args, **kwargs))
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1360, in add_q
clause, _ = self._add_q(q_object, self.used_aliases)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1379, in _add_q
child_clause, needed_inner = self.build_filter(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1321, in build_filter
condition = self.build_lookup(lookups, col, value)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1167, in build_lookup
lookup = lookup_class(lhs, rhs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/lookups.py", line 24, in __init__
self.rhs = self.get_prep_lookup()
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/fields/related_lookups.py", line 117, in get_prep_lookup
self.rhs = target_field.get_prep_value(self.rhs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/fields/__init__.py", line 1776, in get_prep_value
raise e.__class__(
Exception Type: TypeError at /api/db/v0/tables/12454/records/
Exception Value: Field 'id' expected a number but got <django.contrib.auth.models.AnonymousUser object at 0x7fe07d769820>.
```
</details>
CC @mathemancer @kgodey
I'm putting this in the First Release milestone, but only because I think someone from the backend team should look at it and then decide how to better prioritize it.
| True | Auth problem with the records endpoint after session expiry - - After my session on the staging site expired, I pressed the "Refresh" UI button within the Table Page.
- The columns and constraints endpoints responded appropriately with an HTTP 403 and the following body:
```json
[
{
"code": 4003,
"message": "Authentication credentials were not provided.",
"details": { "exception": "Authentication credentials were not provided." }
}
]
```
- But the records endpoint gave me an HTTP 500 and the following Django error
<details>
<summary>Traceback</summary>
```
Environment:
Request Method: GET
Request URL: https://staging.mathesar.org/api/db/v0/tables/12454/records/?limit=500&offset=0&order_by=%5B%7B%22field%22%3A79900%2C%22direction%22%3A%22asc%22%7D%5D&filter=%7B%22and%22%3A%5B%7B%22equal%22%3A%5B%7B%22column_id%22%3A%5B62229%5D%7D%2C%7B%22literal%22%3A%5B8%5D%7D%5D%7D%2C%7B%22lesser%22%3A%5B%7B%22column_id%22%3A%5B62697%5D%7D%2C%7B%22literal%22%3A%5B%22100%22%5D%7D%5D%7D%5D%7D
Django Version: 3.1.14
Python Version: 3.9.2
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
'mathesar.middleware.CursorClosedHandlerMiddleware',
'mathesar.middleware.PasswordChangeNeededMiddleware',
'django_userforeignkey.middleware.UserForeignKeyMiddleware',
'django_request_cache.middleware.RequestCacheMiddleware']
Traceback (most recent call last):
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/fields/__init__.py", line 1774, in get_prep_value
return int(value)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/contrib/auth/models.py", line 414, in __int__
raise TypeError('Cannot cast AnonymousUser to int. Are you trying to use it in place of User?')
The above exception (Cannot cast AnonymousUser to int. Are you trying to use it in place of User?) was the direct cause of the following exception:
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/var/www/staging.mathesar.org/mathesar/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler
raise exc
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 497, in dispatch
self.initial(request, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 415, in initial
self.check_permissions(request)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_framework/views.py", line 332, in check_permissions
if not permission.has_permission(request, self):
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 53, in has_permission
allowed = self._evaluate_statements(statements, request, view, action)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 101, in _evaluate_statements
matched = self._get_statements_matching_conditions(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 235, in _get_statements_matching_conditions
passed = bool(boolExpr.parseString(condition)[0])
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/parsing.py", line 32, in __bool__
return self.evalop(bool(a) for a in self.args)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/parsing.py", line 32, in <genexpr>
return self.evalop(bool(a) for a in self.args)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/parsing.py", line 14, in __bool__
return self.check_condition_fn(self.label)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 218, in <lambda>
check_cond_fn = lambda cond: self._check_condition(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/rest_access_policy/access_policy.py", line 263, in _check_condition
result = method(request, view, action)
File "/var/www/staging.mathesar.org/mathesar/mathesar/api/db/permissions/records.py", line 37, in is_table_viewer
is_schema_viewer = SchemaRole.objects.filter(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/query.py", line 942, in filter
return self._filter_or_exclude(False, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/query.py", line 962, in _filter_or_exclude
clone._filter_or_exclude_inplace(negate, *args, **kwargs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/query.py", line 969, in _filter_or_exclude_inplace
self._query.add_q(Q(*args, **kwargs))
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1360, in add_q
clause, _ = self._add_q(q_object, self.used_aliases)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1379, in _add_q
child_clause, needed_inner = self.build_filter(
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1321, in build_filter
condition = self.build_lookup(lookups, col, value)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/sql/query.py", line 1167, in build_lookup
lookup = lookup_class(lhs, rhs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/lookups.py", line 24, in __init__
self.rhs = self.get_prep_lookup()
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/fields/related_lookups.py", line 117, in get_prep_lookup
self.rhs = target_field.get_prep_value(self.rhs)
File "/opt/virtualenvs/staging-mathesar/lib/python3.9/site-packages/django/db/models/fields/__init__.py", line 1776, in get_prep_value
raise e.__class__(
Exception Type: TypeError at /api/db/v0/tables/12454/records/
Exception Value: Field 'id' expected a number but got <django.contrib.auth.models.AnonymousUser object at 0x7fe07d769820>.
```
</details>
CC @mathemancer @kgodey
I'm putting this in the First Release milestone, but only because I think someone from the backend team should look at it and then decide how to better prioritize it.
| main | auth problem with the records endpoint after session expiry after my session on the staging site expired i pressed the refresh ui button within the table page the columns and constraints endpoints responded appropriately with an http and the following body json code message authentication credentials were not provided details exception authentication credentials were not provided but the records endpoint gave me an http and the following django error traceback environment request method get request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware mathesar middleware cursorclosedhandlermiddleware mathesar middleware passwordchangeneededmiddleware django userforeignkey middleware userforeignkeymiddleware django request cache middleware requestcachemiddleware traceback most recent call last file opt virtualenvs staging mathesar lib site packages django db models fields init py line in get prep value return int value file opt virtualenvs staging mathesar lib site packages django contrib auth models py line in int raise typeerror cannot cast anonymoususer to int are you trying to use it in place of user the above exception cannot cast anonymoususer to int are you trying to use it in place of user was the direct cause of the following exception file opt virtualenvs staging mathesar lib site packages django core handlers exception py line in inner response get response request file opt virtualenvs staging mathesar lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file opt virtualenvs staging mathesar lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file opt virtualenvs staging mathesar lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file opt virtualenvs staging mathesar lib site packages rest framework views py line in dispatch response self handle exception exc file opt virtualenvs staging mathesar lib site packages rest framework views py line in handle exception response exception handler exc context file var www staging mathesar org mathesar mathesar exception handlers py line in mathesar exception handler raise exc file opt virtualenvs staging mathesar lib site packages rest framework views py line in dispatch self initial request args kwargs file opt virtualenvs staging mathesar lib site packages rest framework views py line in initial self check permissions request file opt virtualenvs staging mathesar lib site packages rest framework views py line in check permissions if not permission has permission request self file opt virtualenvs staging mathesar lib site packages rest access policy access policy py line in has permission allowed self evaluate statements statements request view action file opt virtualenvs staging mathesar lib site packages rest access policy access policy py line in evaluate statements matched self get statements matching conditions file opt virtualenvs staging mathesar lib site packages rest access policy access policy py line in get statements matching conditions passed bool boolexpr parsestring condition file opt virtualenvs staging mathesar lib site packages rest access policy parsing py line in bool return self evalop bool a for a in self args file opt virtualenvs staging mathesar lib site packages rest access policy parsing py line in return self evalop bool a for a in self args file opt virtualenvs staging mathesar lib site packages rest access policy parsing py line in bool return self check condition fn self label file opt virtualenvs staging mathesar lib site packages rest access policy access policy py line in check cond fn lambda cond self check condition file opt virtualenvs staging mathesar lib site packages rest access policy access policy py line in check condition result method request view action file var www staging mathesar org mathesar mathesar api db permissions records py line in is table viewer is schema viewer schemarole objects filter file opt virtualenvs staging mathesar lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file opt virtualenvs staging mathesar lib site packages django db models query py line in filter return self filter or exclude false args kwargs file opt virtualenvs staging mathesar lib site packages django db models query py line in filter or exclude clone filter or exclude inplace negate args kwargs file opt virtualenvs staging mathesar lib site packages django db models query py line in filter or exclude inplace self query add q q args kwargs file opt virtualenvs staging mathesar lib site packages django db models sql query py line in add q clause self add q q object self used aliases file opt virtualenvs staging mathesar lib site packages django db models sql query py line in add q child clause needed inner self build filter file opt virtualenvs staging mathesar lib site packages django db models sql query py line in build filter condition self build lookup lookups col value file opt virtualenvs staging mathesar lib site packages django db models sql query py line in build lookup lookup lookup class lhs rhs file opt virtualenvs staging mathesar lib site packages django db models lookups py line in init self rhs self get prep lookup file opt virtualenvs staging mathesar lib site packages django db models fields related lookups py line in get prep lookup self rhs target field get prep value self rhs file opt virtualenvs staging mathesar lib site packages django db models fields init py line in get prep value raise e class exception type typeerror at api db tables records exception value field id expected a number but got cc mathemancer kgodey i m putting this in the first release milestone but only because i think someone from the backend team should look at it and then decide how to better prioritize it | 1 |
5,823 | 30,835,004,157 | IssuesEvent | 2023-08-02 06:38:36 | jupyter-naas/awesome-notebooks | https://api.github.com/repos/jupyter-naas/awesome-notebooks | opened | Advertools - Check website pages status code | templates maintainer | This notebook checks the status code of website pages using Advertools. It is usefull for organizations to quickly check the status of their website pages.
| True | Advertools - Check website pages status code - This notebook checks the status code of website pages using Advertools. It is usefull for organizations to quickly check the status of their website pages.
| main | advertools check website pages status code this notebook checks the status code of website pages using advertools it is usefull for organizations to quickly check the status of their website pages | 1 |
8,642 | 8,366,791,023 | IssuesEvent | 2018-10-04 10:09:39 | prometheus/prometheus | https://api.github.com/repos/prometheus/prometheus | closed | EC2 service discovery panics when OwnerID is nil | component/service discovery kind/bug | _This only affects installations specifying the `endpoint` parameter in the EC2 SD configuration which means it doesn't concern users on AWS EC2._
https://github.com/prometheus/prometheus/blob/1c89984778fa78eda1b89d7fc2d30e025c7d3b9a/discovery/ec2/ec2.go#L243-L246
When `r.OwnerId` is `nil`, the server will crash. | 1.0 | EC2 service discovery panics when OwnerID is nil - _This only affects installations specifying the `endpoint` parameter in the EC2 SD configuration which means it doesn't concern users on AWS EC2._
https://github.com/prometheus/prometheus/blob/1c89984778fa78eda1b89d7fc2d30e025c7d3b9a/discovery/ec2/ec2.go#L243-L246
When `r.OwnerId` is `nil`, the server will crash. | non_main | service discovery panics when ownerid is nil this only affects installations specifying the endpoint parameter in the sd configuration which means it doesn t concern users on aws when r ownerid is nil the server will crash | 0 |
1,810 | 6,576,169,928 | IssuesEvent | 2017-09-11 18:47:08 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | add_host hosts doesn't inherit host_vars from a vars plugin | affects_2.1 needs_info waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report (apologies for opening this in the wrong repo -- https://github.com/ansible/ansible/issues/17544)
##### COMPONENT NAME
add_host
##### ANSIBLE VERSION
ansible 2.1.0.0
##### CONFIGURATION
ansible.cfg
``` ini
[defaults]
vars_plugins = vars_plugins
```
##### OS / ENVIRONMENT
N/A
##### SUMMARY
When adding hosts with `add_host` anything set via vars_plugins through for instance `host.set_variable` doesn't seem to be added.
##### STEPS TO REPRODUCE
playbook.yml
``` yams
- hosts: localhost
gather_facts: false
tasks:
- add_host: name="abc" foo="bar"
- debug: var=hostvars['abc']
```
vars_plugins/test.py
``` python
class VarsModule(object):
def __init__(self, inventory):
""" constructor """
self.inventory = inventory
self.inventory_basedir = inventory.basedir()
def run(self, host, vault_password=None):
return {}
def get_host_vars(self, host, vault_password=None):
host.set_variable('foo', 'should be baz')
return {}
def get_group_vars(self, group, vault_password=None):
return {}
```
##### EXPECTED RESULTS
Debug output should display an overridden "foo" variable (I tried adding other empty variables too)
##### ACTUAL RESULTS
It's unchanged:
``` json
{
"hostvars['abc']": {
"ansible_check_mode": false,
"ansible_version": {
"full": "2.1.0.0",
"major": 2,
"minor": 1,
"revision": 0,
"string": "2.1.0.0"
},
"foo": "bar",
"group_names": [],
"groups": {
"all": [
"localhost",
"abc"
],
"ungrouped": [
"localhost"
]
},
"inventory_dir": null,
"inventory_file": null,
"inventory_hostname": "abc",
"inventory_hostname_short": "abc",
"omit": "__omit_place_holder__62074bbe1679ad30f57196d92a71186467341ed9",
"playbook_dir": "/Users/jbergstroem/Work/ansible-plugin-foo"
}
}
```
| True | add_host hosts doesn't inherit host_vars from a vars plugin - ##### ISSUE TYPE
- Bug Report (apologies for opening this in the wrong repo -- https://github.com/ansible/ansible/issues/17544)
##### COMPONENT NAME
add_host
##### ANSIBLE VERSION
ansible 2.1.0.0
##### CONFIGURATION
ansible.cfg
``` ini
[defaults]
vars_plugins = vars_plugins
```
##### OS / ENVIRONMENT
N/A
##### SUMMARY
When adding hosts with `add_host` anything set via vars_plugins through for instance `host.set_variable` doesn't seem to be added.
##### STEPS TO REPRODUCE
playbook.yml
``` yams
- hosts: localhost
gather_facts: false
tasks:
- add_host: name="abc" foo="bar"
- debug: var=hostvars['abc']
```
vars_plugins/test.py
``` python
class VarsModule(object):
def __init__(self, inventory):
""" constructor """
self.inventory = inventory
self.inventory_basedir = inventory.basedir()
def run(self, host, vault_password=None):
return {}
def get_host_vars(self, host, vault_password=None):
host.set_variable('foo', 'should be baz')
return {}
def get_group_vars(self, group, vault_password=None):
return {}
```
##### EXPECTED RESULTS
Debug output should display an overridden "foo" variable (I tried adding other empty variables too)
##### ACTUAL RESULTS
It's unchanged:
``` json
{
"hostvars['abc']": {
"ansible_check_mode": false,
"ansible_version": {
"full": "2.1.0.0",
"major": 2,
"minor": 1,
"revision": 0,
"string": "2.1.0.0"
},
"foo": "bar",
"group_names": [],
"groups": {
"all": [
"localhost",
"abc"
],
"ungrouped": [
"localhost"
]
},
"inventory_dir": null,
"inventory_file": null,
"inventory_hostname": "abc",
"inventory_hostname_short": "abc",
"omit": "__omit_place_holder__62074bbe1679ad30f57196d92a71186467341ed9",
"playbook_dir": "/Users/jbergstroem/Work/ansible-plugin-foo"
}
}
```
| main | add host hosts doesn t inherit host vars from a vars plugin issue type bug report apologies for opening this in the wrong repo component name add host ansible version ansible configuration ansible cfg ini vars plugins vars plugins os environment n a summary when adding hosts with add host anything set via vars plugins through for instance host set variable doesn t seem to be added steps to reproduce playbook yml yams hosts localhost gather facts false tasks add host name abc foo bar debug var hostvars vars plugins test py python class varsmodule object def init self inventory constructor self inventory inventory self inventory basedir inventory basedir def run self host vault password none return def get host vars self host vault password none host set variable foo should be baz return def get group vars self group vault password none return expected results debug output should display an overridden foo variable i tried adding other empty variables too actual results it s unchanged json hostvars ansible check mode false ansible version full major minor revision string foo bar group names groups all localhost abc ungrouped localhost inventory dir null inventory file null inventory hostname abc inventory hostname short abc omit omit place holder playbook dir users jbergstroem work ansible plugin foo | 1 |
26,301 | 12,397,987,071 | IssuesEvent | 2020-05-21 00:22:34 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Preview or GA? | Pri1 assigned-to-author container-service/svc doc-enhancement triaged | 'Control egress traffic' as described in https://docs.microsoft.com/en-us/azure/aks/limit-egress-traffic#egress-traffic-overview seems to depend on this feature. Since 'limit egress trafic' is GA https://azure.microsoft.com/en-us/updates/egress-lockdown-in-azure-kubernetes-service-aks-is-now-generally-available/ I would assume egress-outboundtype is also GA, unlike what's indicated in this page.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7d721081-0c1c-abf3-a398-386a718446d0
* Version Independent ID: 92b5504d-e25a-be94-e5e5-151083a70607
* Content: [Customize user-defined routes (UDR) in Azure Kubernetes Service (AKS) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/egress-outboundtype)
* Content Source: [articles/aks/egress-outboundtype.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/egress-outboundtype.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | 1.0 | Preview or GA? - 'Control egress traffic' as described in https://docs.microsoft.com/en-us/azure/aks/limit-egress-traffic#egress-traffic-overview seems to depend on this feature. Since 'limit egress trafic' is GA https://azure.microsoft.com/en-us/updates/egress-lockdown-in-azure-kubernetes-service-aks-is-now-generally-available/ I would assume egress-outboundtype is also GA, unlike what's indicated in this page.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7d721081-0c1c-abf3-a398-386a718446d0
* Version Independent ID: 92b5504d-e25a-be94-e5e5-151083a70607
* Content: [Customize user-defined routes (UDR) in Azure Kubernetes Service (AKS) - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/egress-outboundtype)
* Content Source: [articles/aks/egress-outboundtype.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/egress-outboundtype.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | non_main | preview or ga control egress traffic as described in seems to depend on this feature since limit egress trafic is ga i would assume egress outboundtype is also ga unlike what s indicated in this page document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned | 0 |
242,249 | 20,207,011,433 | IssuesEvent | 2022-02-11 21:45:04 | nih-cfde/submission-workflow | https://api.github.com/repos/nih-cfde/submission-workflow | closed | Cancelled submissions don't update/Can't cancel submissions | bug Testing | I had a submission hang, and cancelled it in globus. This triggered an email from globus telling me the submission cancelled, but the submission tool didn't seem to update. The tool continues to say:
```
(cfdesubmit) (base) amandas-mbp-2:~ amanda$ cfde-submit status
Running on service 'staging'
Running on service 'staging'
Status of CFDE Submission (Flow ID 6000961b-3bf1-4acc-a2c5-92e54159480c)
This instance ID: 22ef6868-1a4a-4401-8782-e8ff5d712d2b
This Flow is still in progress.
Current Flow Step: TransferData
```
even after several minutes/several attempts | 1.0 | Cancelled submissions don't update/Can't cancel submissions - I had a submission hang, and cancelled it in globus. This triggered an email from globus telling me the submission cancelled, but the submission tool didn't seem to update. The tool continues to say:
```
(cfdesubmit) (base) amandas-mbp-2:~ amanda$ cfde-submit status
Running on service 'staging'
Running on service 'staging'
Status of CFDE Submission (Flow ID 6000961b-3bf1-4acc-a2c5-92e54159480c)
This instance ID: 22ef6868-1a4a-4401-8782-e8ff5d712d2b
This Flow is still in progress.
Current Flow Step: TransferData
```
even after several minutes/several attempts | non_main | cancelled submissions don t update can t cancel submissions i had a submission hang and cancelled it in globus this triggered an email from globus telling me the submission cancelled but the submission tool didn t seem to update the tool continues to say cfdesubmit base amandas mbp amanda cfde submit status running on service staging running on service staging status of cfde submission flow id this instance id this flow is still in progress current flow step transferdata even after several minutes several attempts | 0 |
818 | 4,441,895,799 | IssuesEvent | 2016-08-19 11:13:28 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | unarchive: no support for .bz2 and .gz files (only tar.* and zip) | bug_report docs_report waiting_on_maintainer | ##### Issue Type:
- Documentation Report
##### Plugin Name:
unarchive
##### Ansible Version:
ansible 2.0.0.2
##### Ansible Configuration:
Default. No changes.
##### Environment:
N/A
##### Summary:
unarchive module only supports tar.* and .zip files, but the documentation does not mention it and the error message is utterly unhelpful. If you try to use the module with .bz2 or .gz file (no tar), the error message says "Failed to find handler for xxxx. Make sure the required command to extract the file is installed.". The commands are present, it's simply the module which does not support extracting those files.
Either the documentation should be amended (currently it indicates that almost any package can be extracted) or the support for more (commonly used) archivers added.
##### Steps To Reproduce:
1. Compress any file using `bzip2 -c`
2. Try using the `unarchive` module with this file.
##### Expected Results:
The file would be extracted, similar to running `bzip2 -d`.
##### Actual Results:
```
fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"backup": null, "content": null, "copy": false, "creates": null, "delimiter": null, "dest": "/tmp/SOURCES/", "directory_mode": null, "follow": false, "force": null, "group": null, "list_files": false, "mode": null, "original_basename": "xxx.gz", "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": "/tmp/SOURCES/xxxl.gz"}}, "msg": "Failed to find handler for \"/tmp/SOURCES/xxx.gz\". Make sure the required command to extract the file is installed."}
``` | True | unarchive: no support for .bz2 and .gz files (only tar.* and zip) - ##### Issue Type:
- Documentation Report
##### Plugin Name:
unarchive
##### Ansible Version:
ansible 2.0.0.2
##### Ansible Configuration:
Default. No changes.
##### Environment:
N/A
##### Summary:
unarchive module only supports tar.* and .zip files, but the documentation does not mention it and the error message is utterly unhelpful. If you try to use the module with .bz2 or .gz file (no tar), the error message says "Failed to find handler for xxxx. Make sure the required command to extract the file is installed.". The commands are present, it's simply the module which does not support extracting those files.
Either the documentation should be amended (currently it indicates that almost any package can be extracted) or the support for more (commonly used) archivers added.
##### Steps To Reproduce:
1. Compress any file using `bzip2 -c`
2. Try using the `unarchive` module with this file.
##### Expected Results:
The file would be extracted, similar to running `bzip2 -d`.
##### Actual Results:
```
fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"backup": null, "content": null, "copy": false, "creates": null, "delimiter": null, "dest": "/tmp/SOURCES/", "directory_mode": null, "follow": false, "force": null, "group": null, "list_files": false, "mode": null, "original_basename": "xxx.gz", "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": "/tmp/SOURCES/xxxl.gz"}}, "msg": "Failed to find handler for \"/tmp/SOURCES/xxx.gz\". Make sure the required command to extract the file is installed."}
``` | main | unarchive no support for and gz files only tar and zip issue type documentation report plugin name unarchive ansible version ansible ansible configuration default no changes environment n a summary unarchive module only supports tar and zip files but the documentation does not mention it and the error message is utterly unhelpful if you try to use the module with or gz file no tar the error message says failed to find handler for xxxx make sure the required command to extract the file is installed the commands are present it s simply the module which does not support extracting those files either the documentation should be amended currently it indicates that almost any package can be extracted or the support for more commonly used archivers added steps to reproduce compress any file using c try using the unarchive module with this file expected results the file would be extracted similar to running d actual results fatal failed changed false failed true invocation module args backup null content null copy false creates null delimiter null dest tmp sources directory mode null follow false force null group null list files false mode null original basename xxx gz owner null regexp null remote src null selevel null serole null setype null seuser null src tmp sources xxxl gz msg failed to find handler for tmp sources xxx gz make sure the required command to extract the file is installed | 1 |
1,567 | 6,572,324,499 | IssuesEvent | 2017-09-11 01:23:24 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | npm: module crashes du to malformed npm output | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### ANSIBLE VERSION
```
ansible 2.1.0.0
```
##### CONFIGURATION
ø
##### OS / ENVIRONMENT
Debian Jessie
##### SUMMARY
`ansible_module_npm.py` raises an exception.
##### STEPS TO REPRODUCE
I'm just having this in my playbook:
```
- name: Install requirements from npm
npm: name={{item}} global=yes state=latest
with_items:
- hogan.js
- bower
- gulp
- cordova
- jscs
- jshint
- ionic
```
##### EXPECTED RESULTS
Some modules installs successfully, but other make npm.py raises an exception.
##### ACTUAL RESULTS
```
<dev4> ESTABLISH SSH CONNECTION FOR USER: root
<dev4> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r dev4 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984 `" && echo ansible-tmp-1468397722.47-174060488007984="` echo $HOME/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984 `" ) && sleep 0'"'"''
<dev4> PUT /tmp/tmpU9v9LP TO /root/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984/npm
<dev4.eeple.fr> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r '[dev4.eeple.fr]'
<dev4> ESTABLISH SSH CONNECTION FOR USER: root
<dev4> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r -tt dev4.eeple.fr '/bin/sh -c '"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984/npm; rm -rf "/root/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984/" > /dev/null 2>&1 && sleep 0'"'"''
failed: [dev4] (item=jscs) => {"failed": true, "invocation": {"module_name": "npm"}, "item": "jscs", "module_stderr": "", "module_stdout": "
(manually redacted in github issue for readability:)
Traceback (most recent call last):
File "/tmp/ansible_1jeeYZ/ansible_module_npm.py", line 269, in <module>
main()
File "/tmp/ansible_1jeeYZ/ansible_module_npm.py", line 255, in main
outdated = npm.list_outdated()
File "/tmp/ansible_1jeeYZ/ansible_module_npm.py", line 206, in list_outdated
pkg, other = re.split('\\s|@', dep, 1)
ValueError: need more than 1 value to unpack
", "msg": "MODULE FAILURE", "parsed": false}
```
I inspected what happen exactly, and ansible runs `/usr/bin/npm outdated --global gulp` which gives:
```
Package Current Wanted Latest Location
gulp 3.9.0 3.9.1 3.9.1 /usr/local/lib > gulp
gulp 3.8.8 3.8.8 3.9.1 /usr/local/lib > ionic > gulp
undefined
```
So I opened an issue on npm too: https://github.com/npm/npm/issues/13364 but I think the module can be enhanced and continue working normally if npm prints garbage. A warning may be emmited but it may be possible to continue and install the package.
| True | npm: module crashes du to malformed npm output - ##### ISSUE TYPE
- Bug Report
##### ANSIBLE VERSION
```
ansible 2.1.0.0
```
##### CONFIGURATION
ø
##### OS / ENVIRONMENT
Debian Jessie
##### SUMMARY
`ansible_module_npm.py` raises an exception.
##### STEPS TO REPRODUCE
I'm just having this in my playbook:
```
- name: Install requirements from npm
npm: name={{item}} global=yes state=latest
with_items:
- hogan.js
- bower
- gulp
- cordova
- jscs
- jshint
- ionic
```
##### EXPECTED RESULTS
Some modules installs successfully, but other make npm.py raises an exception.
##### ACTUAL RESULTS
```
<dev4> ESTABLISH SSH CONNECTION FOR USER: root
<dev4> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r dev4 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984 `" && echo ansible-tmp-1468397722.47-174060488007984="` echo $HOME/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984 `" ) && sleep 0'"'"''
<dev4> PUT /tmp/tmpU9v9LP TO /root/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984/npm
<dev4.eeple.fr> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r '[dev4.eeple.fr]'
<dev4> ESTABLISH SSH CONNECTION FOR USER: root
<dev4> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r -tt dev4.eeple.fr '/bin/sh -c '"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984/npm; rm -rf "/root/.ansible/tmp/ansible-tmp-1468397722.47-174060488007984/" > /dev/null 2>&1 && sleep 0'"'"''
failed: [dev4] (item=jscs) => {"failed": true, "invocation": {"module_name": "npm"}, "item": "jscs", "module_stderr": "", "module_stdout": "
(manually redacted in github issue for readability:)
Traceback (most recent call last):
File "/tmp/ansible_1jeeYZ/ansible_module_npm.py", line 269, in <module>
main()
File "/tmp/ansible_1jeeYZ/ansible_module_npm.py", line 255, in main
outdated = npm.list_outdated()
File "/tmp/ansible_1jeeYZ/ansible_module_npm.py", line 206, in list_outdated
pkg, other = re.split('\\s|@', dep, 1)
ValueError: need more than 1 value to unpack
", "msg": "MODULE FAILURE", "parsed": false}
```
I inspected what happen exactly, and ansible runs `/usr/bin/npm outdated --global gulp` which gives:
```
Package Current Wanted Latest Location
gulp 3.9.0 3.9.1 3.9.1 /usr/local/lib > gulp
gulp 3.8.8 3.8.8 3.9.1 /usr/local/lib > ionic > gulp
undefined
```
So I opened an issue on npm too: https://github.com/npm/npm/issues/13364 but I think the module can be enhanced and continue working normally if npm prints garbage. A warning may be emmited but it may be possible to continue and install the package.
| main | npm module crashes du to malformed npm output issue type bug report ansible version ansible configuration ø os environment debian jessie summary ansible module npm py raises an exception steps to reproduce i m just having this in my playbook name install requirements from npm npm name item global yes state latest with items hogan js bower gulp cordova jscs jshint ionic expected results some modules installs successfully but other make npm py raises an exception actual results establish ssh connection for user root ssh exec ssh c q o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home julien ansible cp ansible ssh h p r bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to root ansible tmp ansible tmp npm ssh exec sftp b c o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home julien ansible cp ansible ssh h p r establish ssh connection for user root ssh exec ssh c q o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home julien ansible cp ansible ssh h p r tt eeple fr bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python root ansible tmp ansible tmp npm rm rf root ansible tmp ansible tmp dev null sleep failed item jscs failed true invocation module name npm item jscs module stderr module stdout manually redacted in github issue for readability traceback most recent call last file tmp ansible ansible module npm py line in main file tmp ansible ansible module npm py line in main outdated npm list outdated file tmp ansible ansible module npm py line in list outdated pkg other re split s dep valueerror need more than value to unpack msg module failure parsed false i inspected what happen exactly and ansible runs usr bin npm outdated global gulp which gives package current wanted latest location gulp usr local lib gulp gulp usr local lib ionic gulp undefined so i opened an issue on npm too but i think the module can be enhanced and continue working normally if npm prints garbage a warning may be emmited but it may be possible to continue and install the package | 1 |
3,562 | 14,258,373,855 | IssuesEvent | 2020-11-20 06:08:30 | SethMMorton/natsort | https://api.github.com/repos/SethMMorton/natsort | closed | Better support for unicode numbers | good first issue help wanted maintainability | **Describe the feature or enhancement**
There should be some way to determine all the unicode numbers at start-up with zero runtime cost that does not require releasing a new `natsort` version.
**Provide a concrete example of how the feature or enhancement will improve `natsort`**
Each time a new Python version is released, `natsort` needs to make an update so that unit tests do not fail. Often, users spot this before I do (#104, #119). The reason I did it this way is because determining this at import time introduces a noticeable slowdown that I am unwilling to make my users incur.
**I am looking for ideas/help on how to solve this so that I do not hurt user experience while simultaneously lowering maintenance burden**.
| True | Better support for unicode numbers - **Describe the feature or enhancement**
There should be some way to determine all the unicode numbers at start-up with zero runtime cost that does not require releasing a new `natsort` version.
**Provide a concrete example of how the feature or enhancement will improve `natsort`**
Each time a new Python version is released, `natsort` needs to make an update so that unit tests do not fail. Often, users spot this before I do (#104, #119). The reason I did it this way is because determining this at import time introduces a noticeable slowdown that I am unwilling to make my users incur.
**I am looking for ideas/help on how to solve this so that I do not hurt user experience while simultaneously lowering maintenance burden**.
| main | better support for unicode numbers describe the feature or enhancement there should be some way to determine all the unicode numbers at start up with zero runtime cost that does not require releasing a new natsort version provide a concrete example of how the feature or enhancement will improve natsort each time a new python version is released natsort needs to make an update so that unit tests do not fail often users spot this before i do the reason i did it this way is because determining this at import time introduces a noticeable slowdown that i am unwilling to make my users incur i am looking for ideas help on how to solve this so that i do not hurt user experience while simultaneously lowering maintenance burden | 1 |
430,906 | 12,467,912,032 | IssuesEvent | 2020-05-28 17:54:31 | square/bazel_maven_repository | https://api.github.com/repos/square/bazel_maven_repository | opened | Make classes.jar more IDE friendly | Priority-2 Type-enhancement | Right now, you end up with a lot of "classes-<somekey>.jar" in your external workspace list in intellij. Renaming classes.jar to something more verbose can help disambiguate them in the external libraries list. Either artifactId-version-classes.jar (e.g. guava-18.0-classes.jar) would improve their visual appearance. possibly even adding some portion fo the group id to them. Definitely should do the former, need to consider the latter (adding group disambiguation). | 1.0 | Make classes.jar more IDE friendly - Right now, you end up with a lot of "classes-<somekey>.jar" in your external workspace list in intellij. Renaming classes.jar to something more verbose can help disambiguate them in the external libraries list. Either artifactId-version-classes.jar (e.g. guava-18.0-classes.jar) would improve their visual appearance. possibly even adding some portion fo the group id to them. Definitely should do the former, need to consider the latter (adding group disambiguation). | non_main | make classes jar more ide friendly right now you end up with a lot of classes jar in your external workspace list in intellij renaming classes jar to something more verbose can help disambiguate them in the external libraries list either artifactid version classes jar e g guava classes jar would improve their visual appearance possibly even adding some portion fo the group id to them definitely should do the former need to consider the latter adding group disambiguation | 0 |
1,048 | 4,861,978,538 | IssuesEvent | 2016-11-14 10:43:59 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | blockinfile replaces the block instead of inserting it after EOF | affects_2.3 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
blockinfile
##### ANSIBLE VERSION
```
ansible 2.3.0
commit 20161109.65f019f
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.10 4.8
##### SUMMARY
When blockinfile is used in a loop to append multiple blocks at the end of the file, it replaces previous block instead with the current one so we end up with only the last block in the file.
##### STEPS TO REPRODUCE:
Trying to append all blocks (with the same marker) at the end of file with:
```
- name: Deleting "{{ item }}" when local file already exists
local_action: file dest="{{ dest_file }}" state=absent
- name: Saving "{{ item }}" into local file
blockinfile:
dest: "{{ dest_file }}"
create: yes
block: '{{ stdout_item }}'
marker: "<--- {mark} {{item}} --->"
insertafter: EOF
with_items: "{{ table.stdout }}"
loop_control:
loop_var: stdout_item
```
leads to:
```
TASK [save_table : Saving "IPv4_Routing_Table" into local file] *************************************************************************************************************
changed: [ASAv1] => (item=
IP routing table maximum-paths is 8
Route Source Networks Subnets Replicates Overhead Memory (bytes)
connected 0 14 0 1232 4032
static 1 0 0 88 288
bgp 12 0 28 0 4664 8064
External: 28 Internal: 0 Local: 0
internal 5 4440
Total 6 42 0 5984 16824
) => {"changed": true, "msg": "Block inserted", "stdout_item": "\nIP routing table maximum-paths is 8\nRoute Source Networks Subnets Replicates Overhead Memory (bytes)\nconnected 0 14 0 1232 4032 \nstatic 1 0 0 88 288 \nbgp 12 0 28 0 4664 8064 \n External: 28 Internal: 0 Local: 0\ninternal 5 4440 \nTotal 6 42 0 5984 16824 \n\n"}
changed: [ASAv1] => (item=
Codes: L - local, C - connected, S - static, R - RIP, M - mobile, B - BGP
D - EIGRP, EX - EIGRP external, O - OSPF, IA - OSPF inter area
N1 - OSPF NSSA external type 1, N2 - OSPF NSSA external type 2
E1 - OSPF external type 1, E2 - OSPF external type 2
i - IS-IS, su - IS-IS summary, L1 - IS-IS level-1, L2 - IS-IS level-2
ia - IS-IS inter area, * - candidate default, U - per-user static route
o - ODR, P - periodic downloaded static route, + - replicated route
Gateway of last resort is 192.168.122.1 to network 0.0.0.0
S* 0.0.0.0 0.0.0.0 [1/0] via 192.168.122.1, Bouygues-Telecom
C 4.65.13.0 255.255.255.0 is directly connected, Level-3
...
) => {"changed": true, "msg": "Block inserted", "stdout_item": "\nCodes: L - local, C - connected, S - static, R - RIP, M - mobile, B - BGP\n D - EIGRP, EX - EIGRP external, O - OSPF, IA - OSPF inter area \n N1 - OSPF NSSA external type 1, N2 - OSPF NSSA external type 2\n E1 - OSPF external type 1, E2 - OSPF external type 2\n i - IS-IS, su - IS-IS summary, L1 - IS-IS level-1, L2 - IS-IS level-2\n ia - IS-IS inter area, * - candidate default, U - per-user static route\n o - ODR, P - periodic downloaded static route, + - replicated route\nGateway of last resort is 192.168.122.1 to network 0.0.0.0\n\nS* 0.0.0.0 0.0.0.0 [1/0] via 192.168.122.1, Bouygues-Telecom\nC 4.65.13.0 255.255.255.0 is directly connected, Level-3\n..."}
```
But **in the file, only the second block remains**, the first one has been replaced:
```
<--- BEGIN IPv4_Routing_Table --->
Codes: L - local, C - connected, S - static, R - RIP, M - mobile, B - BGP
D - EIGRP, EX - EIGRP external, O - OSPF, IA - OSPF inter area
N1 - OSPF NSSA external type 1, N2 - OSPF NSSA external type 2
E1 - OSPF external type 1, E2 - OSPF external type 2
i - IS-IS, su - IS-IS summary, L1 - IS-IS level-1, L2 - IS-IS level-2
ia - IS-IS inter area, * - candidate default, U - per-user static route
o - ODR, P - periodic downloaded static route, + - replicated route
Gateway of last resort is 192.168.122.1 to network 0.0.0.0
S* 0.0.0.0 0.0.0.0 [1/0] via 192.168.122.1, Bouygues-Telecom
C 4.65.13.0 255.255.255.0 is directly connected, Level-3
...
<--- END IPv4_Routing_Table --->
```
| True | blockinfile replaces the block instead of inserting it after EOF - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
blockinfile
##### ANSIBLE VERSION
```
ansible 2.3.0
commit 20161109.65f019f
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.10 4.8
##### SUMMARY
When blockinfile is used in a loop to append multiple blocks at the end of the file, it replaces previous block instead with the current one so we end up with only the last block in the file.
##### STEPS TO REPRODUCE:
Trying to append all blocks (with the same marker) at the end of file with:
```
- name: Deleting "{{ item }}" when local file already exists
local_action: file dest="{{ dest_file }}" state=absent
- name: Saving "{{ item }}" into local file
blockinfile:
dest: "{{ dest_file }}"
create: yes
block: '{{ stdout_item }}'
marker: "<--- {mark} {{item}} --->"
insertafter: EOF
with_items: "{{ table.stdout }}"
loop_control:
loop_var: stdout_item
```
leads to:
```
TASK [save_table : Saving "IPv4_Routing_Table" into local file] *************************************************************************************************************
changed: [ASAv1] => (item=
IP routing table maximum-paths is 8
Route Source Networks Subnets Replicates Overhead Memory (bytes)
connected 0 14 0 1232 4032
static 1 0 0 88 288
bgp 12 0 28 0 4664 8064
External: 28 Internal: 0 Local: 0
internal 5 4440
Total 6 42 0 5984 16824
) => {"changed": true, "msg": "Block inserted", "stdout_item": "\nIP routing table maximum-paths is 8\nRoute Source Networks Subnets Replicates Overhead Memory (bytes)\nconnected 0 14 0 1232 4032 \nstatic 1 0 0 88 288 \nbgp 12 0 28 0 4664 8064 \n External: 28 Internal: 0 Local: 0\ninternal 5 4440 \nTotal 6 42 0 5984 16824 \n\n"}
changed: [ASAv1] => (item=
Codes: L - local, C - connected, S - static, R - RIP, M - mobile, B - BGP
D - EIGRP, EX - EIGRP external, O - OSPF, IA - OSPF inter area
N1 - OSPF NSSA external type 1, N2 - OSPF NSSA external type 2
E1 - OSPF external type 1, E2 - OSPF external type 2
i - IS-IS, su - IS-IS summary, L1 - IS-IS level-1, L2 - IS-IS level-2
ia - IS-IS inter area, * - candidate default, U - per-user static route
o - ODR, P - periodic downloaded static route, + - replicated route
Gateway of last resort is 192.168.122.1 to network 0.0.0.0
S* 0.0.0.0 0.0.0.0 [1/0] via 192.168.122.1, Bouygues-Telecom
C 4.65.13.0 255.255.255.0 is directly connected, Level-3
...
) => {"changed": true, "msg": "Block inserted", "stdout_item": "\nCodes: L - local, C - connected, S - static, R - RIP, M - mobile, B - BGP\n D - EIGRP, EX - EIGRP external, O - OSPF, IA - OSPF inter area \n N1 - OSPF NSSA external type 1, N2 - OSPF NSSA external type 2\n E1 - OSPF external type 1, E2 - OSPF external type 2\n i - IS-IS, su - IS-IS summary, L1 - IS-IS level-1, L2 - IS-IS level-2\n ia - IS-IS inter area, * - candidate default, U - per-user static route\n o - ODR, P - periodic downloaded static route, + - replicated route\nGateway of last resort is 192.168.122.1 to network 0.0.0.0\n\nS* 0.0.0.0 0.0.0.0 [1/0] via 192.168.122.1, Bouygues-Telecom\nC 4.65.13.0 255.255.255.0 is directly connected, Level-3\n..."}
```
But **in the file, only the second block remains**, the first one has been replaced:
```
<--- BEGIN IPv4_Routing_Table --->
Codes: L - local, C - connected, S - static, R - RIP, M - mobile, B - BGP
D - EIGRP, EX - EIGRP external, O - OSPF, IA - OSPF inter area
N1 - OSPF NSSA external type 1, N2 - OSPF NSSA external type 2
E1 - OSPF external type 1, E2 - OSPF external type 2
i - IS-IS, su - IS-IS summary, L1 - IS-IS level-1, L2 - IS-IS level-2
ia - IS-IS inter area, * - candidate default, U - per-user static route
o - ODR, P - periodic downloaded static route, + - replicated route
Gateway of last resort is 192.168.122.1 to network 0.0.0.0
S* 0.0.0.0 0.0.0.0 [1/0] via 192.168.122.1, Bouygues-Telecom
C 4.65.13.0 255.255.255.0 is directly connected, Level-3
...
<--- END IPv4_Routing_Table --->
```
| main | blockinfile replaces the block instead of inserting it after eof issue type bug report component name blockinfile ansible version ansible commit config file etc ansible ansible cfg configured module search path default w o overrides configuration inventory hosts gathering explicit roles path home actionmystique program files ubuntu ansible roles roles private role vars yes log path var log ansible log fact caching redis fact caching timeout retry files enabled false os environment host ubuntu summary when blockinfile is used in a loop to append multiple blocks at the end of the file it replaces previous block instead with the current one so we end up with only the last block in the file steps to reproduce trying to append all blocks with the same marker at the end of file with name deleting item when local file already exists local action file dest dest file state absent name saving item into local file blockinfile dest dest file create yes block stdout item marker insertafter eof with items table stdout loop control loop var stdout item leads to task changed item ip routing table maximum paths is route source networks subnets replicates overhead memory bytes connected static bgp external internal local internal total changed true msg block inserted stdout item nip routing table maximum paths is nroute source networks subnets replicates overhead memory bytes nconnected nstatic nbgp n external internal local ninternal ntotal n n changed item codes l local c connected s static r rip m mobile b bgp d eigrp ex eigrp external o ospf ia ospf inter area ospf nssa external type ospf nssa external type ospf external type ospf external type i is is su is is summary is is level is is level ia is is inter area candidate default u per user static route o odr p periodic downloaded static route replicated route gateway of last resort is to network s via bouygues telecom c is directly connected level changed true msg block inserted stdout item ncodes l local c connected s static r rip m mobile b bgp n d eigrp ex eigrp external o ospf ia ospf inter area n ospf nssa external type ospf nssa external type n ospf external type ospf external type n i is is su is is summary is is level is is level n ia is is inter area candidate default u per user static route n o odr p periodic downloaded static route replicated route ngateway of last resort is to network n ns via bouygues telecom nc is directly connected level n but in the file only the second block remains the first one has been replaced codes l local c connected s static r rip m mobile b bgp d eigrp ex eigrp external o ospf ia ospf inter area ospf nssa external type ospf nssa external type ospf external type ospf external type i is is su is is summary is is level is is level ia is is inter area candidate default u per user static route o odr p periodic downloaded static route replicated route gateway of last resort is to network s via bouygues telecom c is directly connected level | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.